var/home/core/zuul-output/0000755000175000017500000000000015146000513014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015146025167015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000337021715146025001020255 0ustar corecore*ikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9Gf䅴l"mv?_eGbuuțx{w7ݭ7֫ce% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿-h QZ*U1|t5wKOؾ{mk b2 ܨ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߾)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[<~6]90}(*T7siv'=k 9Q2@vN ( R['>v*;o57sp$3ncx!>t®W>]tF-iܪ%GYbaRvHa}dkD̶*';ک|s_}8yj,('GrgTZ'U鋊TqOſ * /Ijo!՟8`"j}zӲ$k3jS|C7;A)͎V.r?t\WU1ojjr<~Tq> `=tJ!aݡ=h6Yݭw}?lѹ`f_" J9w4ts7NG GGG]ҡgc⌝M b/Ζlpah E ur C&`XR JcwB~R2EL9j7e\(Uё$׿atyХ?*t5z\+`/ErVQUxMҔ&ۈt.3;eg_O ξL1KiYLizpV:C5/=v-}҅"o ']i=Ob㸵2*3d*mQ%"h+ "f "D(~~moH|E3*46$Ag4aX)Ǜƾ9U Ӆ^};ڲ7J9@ kV%gDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?}=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@8i>rH_HI\:U}UE$J @ٚeZE0(8. ϟ'DrW XWz<%fpG"m%6PGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu̿ܣRF]NHw2kp}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\Mz_s߭AF'is^_&uUm$[[5HI4QCZ5!N&D[uiXk&2Bg&Ս7_/6v_cd쿽d@eU XyX_2z>g8:.⺻h()&nO5YE\1t7aSyFxPV19 ĕi%K"IcB j>Pm[E[^oHmmU̸nG Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"Yg"ź)\3mf|ܔMUiU|Ym! #'ukMmQ9Blm]TO1ba.XW x6ܠ9[v35H;-]Um4mMrW-k#~fؤϋu_j*^Wj^qM `-P?k.@%=X#|ۡb1lKcj$bKv[~"N jS4HOkeF3LPyi︅iWk! cAnxu6<7cp?WN $?X3l(?  'Z! ,Z.maO_Bk/m~ޖ(<qRfR"Au\PmLZ"twpuJ` mvf+T!6Ѓjw1ncuwo':o gSPC=]U҅yY9 &K<-na'Xk,P4+`Þ/lX/bjoFO.= w ?>ȑ3n߿z,t s5Z/ Clo-` z?a~b mzkC zFȏ>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwoj-25Hݻ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@T%W?%TbzK-6cb:XeGhl&0Ɠbb__2++oI~!&-[TWvxZ>4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F onKȌsK+D"̽E/"Icƀsu0,gy(&TI{ U܋N5 l͖h"褁lm *#n/Q!m b0X3i)\IN˭% Y&cKoG w 9pM^WϋQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,+ ] O;*X]Eg,5,uZm8pnglVj!p2֬uT[QyB402|2d5K: `Bcz|Rxxl3{c` 1nhJzQHv?hbºܞz=73qSO0}Dc D]ͺ?jgw07'㤸z YJ\Hb9)S2 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ą~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==qrͅٻҦ62L0ډ"ܺ_z9JNȯ=@oUI y4KVMٞM9$1#HR1(7x]mD@0ngd6#eMy"[ ^Q $[d8  i#i8YlsI!2(ȐP'3ޜb6xo^fmIx nf^Lw>"0(HKkD4<80: M:'֥P!r "Lӓݰ@ 9n# " $fGgKQӦ4}Gn\^=-Y5PI dPN6 Ozځ/פ|5) F[ڣ$2*%&h v%9HN H~Q+oi?&۳)-nqK?2ސv/3,9ҮT9Cef˝49i.2DxatC<8iR/ƬйR֌vN8J"iJ. T>)qaY4ͬlyg "]BvW#99`TegõII kюHLa^c&/H^FFIu`2a$mc Ry+R:LڕDܓ>Y:]t.+|PT6=qWe0NƏw<6o3mv8k vGOfpEOkÈWȤMف lOc;SR&.w,qk>MPs+Xh4iyuGRd֞q鮺]m S{}]U kV0/ŜxtADx"Xh4|;XSxߵă@pE:y]/"(MCG`ʶϊGi+39#gNZYE:Qw9muB`9`LDhs4Ǩ9S`EkM{zB<˙ik; JD;;3!4 2Y.$Dwiu|+lO:k$]ԜYLUҞ6EmH>azʳ/A+ԀZk"f`.,ל{=wh|_qYj5M{K$gv>cDp"'0޽5xCNQ1G2})*'>fC۝'*)"5.E2IeD 2.ZdrN6Uœ=n8D-9޵JKw5ُJ,􋃓ZUꋼ0b1f87GՂ 1t_o}{Mr7KO0Ao-Y*Is\S:JzA(:i!eҎ\,f+,Ąt78~ڋ~?[F^.A'!,iGow3{'YToҝf5ޓ[he>=7S8DGZ@-#]f:Tm?L{F-8G#%.fM8Y='gیl0HڜHLK'Cw#)krWIk<1څ 9abHl:b3LjOq͂Ӥ=u8#E2;|z꽐vɀi^lUt␚ɓW%OVc8|*yI0U=nFGA`IC8p+C:!}Nh,mn>_MGiq'N~|z`|mu}r:"KiyGҪ$& hw#4qn?ܶХfm_Ov^ܶ[6j3ZN9t9ZMMM)I[Rχ/C|W䳮yI3MڼH9iEG&V 'x`u.̀ab7V<*EzfH{]:*6M x-v쳎M'.hO3p-IGh ܆hR ]zi2hB9'S_;I/d0oIU:m/~[*K1QA="D:V&f:{7N>^uU` c/X)mS5KC߄":{H)"%,!3w{"ZWÂk>/F?RJ>FIY*%5Hg}3Ď89؟N/pgÞ tJXB-Gjsٶ 3Gzp؍H|*cyp@\첹,[up`uV,\KCB\qGiW痃[?i?S{eϻl71X:݌>EEly(*SHN:ӫOq{{L$?Q{϶(F_Ej>3mqfΤP-j)H˧&8?a?2xĐ+EV؍x0bv6 fd1^ 2ӎԥ sZR cgu/bn/34'h9Dݥ:U:vV[ 'Mȥ@ەX㧿-p0?Q6 y2XN2_h~Cֆ֙82)=Ȓ7D- V)T? O/VFeUk'7KIT, WeՔ}-66V؅ʹ;T$pZ#@L; ?0]"2v[hׂ'cJ6H4bs+3(@z$.K!#Šj2ݢxK-di +9Hᇷ絻+ O.i2.I+69EVyw8//|~<ëng)P<xͯ~? fp,CǴ_BjDN^5)s('cBh+6ez0)_~zJz"ё`Z&Z![0rGBK 5G~<:H~W>;ٍVnSt%_!BZMMeccBҎÒJH+"ūyR}X~juPp- j\hЪQxchKaS,xS"cV8i8'-sOKB<չw"|{/MC8&%Og3E#O%`N)p#4YUh^ ɨڻ#Ch@(R &Z+<3ݰb/St=&yo|BL,1+t C<ˉvRfQ*e"T:*Dᰤ*~IClz^F6!ܠqK3%$E)~?wy,u'u() C>Gn} t]2_}!1NodI_Bǂ/^8\3m!'(Ֆ5Q&xo 8;'Jbo&XL_ʣ^^"Lq2E3,v1ɢu^}G7Z/qC^'+HDy=\]?d|9i,p?߼=\Ce"|Rݷ Q+=zxB.^Bld.HSntºB4~4]%.i|҂"? ~#ݤ[tfv3Ytck0O ͧ gP\|bЯ݃5H+v}$n;6q.^@Dvܫ46NGZV 9B!JV_K2*<Ɩ9b&2] -Of;M.~o“5 LIa6}_n;ɗBQi'CXt!*< 0U-(pc[{*CqKGp&E]&x!i6١˭K&FCfJ9$ٵQ·BD-]R1#mTRrS [;Zc5qU𢛲@m;Uyfo%pg/cC l 2JZ=0]Sה(*CjaS:p/N6I*Mx8"EȿQa[1 ŶD3u8j`B59qU]ג`upHЍE_fNTU*q%b1! `ʗrǚ8ce){F74~'*{[\M-~#aSmMÉB4NnꝤʇ)b^g`u2?x"8U [JYSk( "nu\h1Yhl~[mhm+F(g +YtyOg ]7n]Q!Mǧę5bR!ȯe |M>8l WIf|\q4|UjA.gr`˱LO{ xr.~l-ɩܿ* %xbcDa.E h Ƶ:R .qɱmu$I8>^QUAZa$1aH_duoI`b#FOf_$0[Ck(OA/$ ^%IA-(Vr]RS1ު"M4A/v1H?>7f/"(o9/1X;[q`/{ 9a>E1XS+!rV%¢EN$q^B^rX؅ z1ǡ~,|I*0>CjỚ,?V~ @VC|J,Ah%ShՉ8Y4czt~$)*527l;V jэy(_,jm$9O+/S`]icm wq\Mڗ$-DJoj*}b,?XAe0bX@ h0w[BU0v']#Vvo !ې:Z%ƶ(fl>'"BgcM_-AnG q\fw$<;u8I\XA-{ύlKwŅjКVhfG8>VƇg1U nASaSK1OOȩ<Q 9UR=Wd<)tc(߯)Y_5>61C( .Jog&_P9&`|8|HdO۷&i"yNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*mMq!R{QWR=oVbmyanUn.Uqsy.?Ws_u3ܸ'AqC_oB㖟E-? k[~Cz iE2/h WJt{=xL@5赠M8GZ,\c EI l1z ɏ,dE>-KjLGgL2h燲 K\!f3KXb^$fH1_G9(ΟRYpŘ-Ŷ֣8N,౬CxAX4xM"5XITd E$Vd VI8ʁu EA>GІ'JƖ'ZXO='kj 3G`mJc,QhZ KNڢw2aK0Z' O 7>*Fx ztH͸FsQ t6Uǩj'RM*ߚԘ!O$ +Jp[qv؆ R!kG,:\nm^ zDpC65M++r0Z[r0d/N#X.?_םh 2r-'OID3d 6RcU0ޮz$q9`尻]Tc\b+s~p"eʰ6Qz\=.!Bjіd0A*Љ&#xu7IAZbbx:C~L :4^`jSؒd 3m%ijFP>TU?!$VA`Rc1wM "U8V!5> =҆#xɮ]U`w۸ہ :Xx+|X!~Pu(ǍUeS@%NբZ.SV! Ž^A12JQ̛rFI_"*lOY>WK-.`oq@>4! /P4Qz5 .+h+0NJ|!N9/|a|>lX9T ҇ :W?*n;fw*B#d[Mȹ4NKLw*Z 'Sa5sy 9VqCb!bSt㒢NCW:F*`Vn-zߞ +8 {mJ697@8sMZXiC U+AW^ܲ'SN1ۀS`+f'r[8ޝ"sդSPW k`u yg[~S [jӋksE.,5 "ka+n!e߭lɹ՚f 2tk)„+A= 2䃔A:G\ ` \vc"Kjm61 Px"3Qc /[?$ PW*3GX lqv-6W&)cX |];C%sXY2%£5@/~+ɇ~t#s1pVʂ4L tׁG- zM2X^"[r>$9{[?- &\j7 )UJwq:ǫf!NRT1D(7.8Qӆ?N+JL0SU$jfˬ!މZ|VA/6ȍ&XLE4atGS1px"S}MFڇ˦NJPYDX%\ꡗwhl}i9f?q>*1ZVƭD:&+ED>M-'[0AG꠭y*s)/=rnl{ӕЩsdLyVIuUI':8^6$ t,O6lb 2ɀ,Eʿk܍fK58$58?DX 4q}ll9WyU D=0.37}:xI/ió21׭ȦS!e^tk28b$dhLXULi7/1-Qk%ƩJ4^plme;u>6nQe UZA *^Vif]>HUd6ƕ̽=&`\g2v%tJ4[7g"zw1|\*& >V! :VNS7[{ݶ%[Ngpn][#&ߓTo_ =2A/a<`~MEp";SB/eWٹj%Jg:lDԒ͢Zu[kWw{7Oй d:[mLة :*mUAN=VήM> jwB}Ԧ\W+[3=`BZWX vZd>t*U8ǖv\}BN`+u6Y3[yBPj|Lc~Ì{rQۻou}q}hn+.{pWEqws]]|/ǫ\}/J.MLmc [ԗWrU}/Ǜ+sYn[ﯾeywyY]]¨Kpx c./m_;ߟRy*4mw6 |ʨOr0 &ve1=VzpO֠7hf 1thiD{q:n%̈#v^ݔ͆!`C}Lu_~ǸTߧcr d9d ljB?KWzzKNf@r3 Zj/I+ϟYSU&+|t><,Bq./J2޲mi^b``Q=dBxV=Vʤލcv4?>HB78NE@Uc8>`TZZ} ?md+P}0/N7`d7%y ٞoIɒci'VMî_UʸbQ4VQex¤^c)RYUIU2ޝkǪ9 #BG"/RJrX1b,|_Ÿq<(GѦyyRS/QxO4n}sgc@8/2,o<%:%oN0/444M__]] ؋@ K/+KsB53csu4 \eqeX0CC{Y\jk^x`m p< n` Ct[BZkU͵dEɳ텶:3UseY;认ƫ@^lbtZbD^`4<}wǓ6|ֿ{<:" ҽ8vy8p;9 /H'wBn>7_~oͿY!ls!>fY%C|j*PÃ[|(Sb~+TY[AH~w-V*KT&PȐۮ*L|6n]n^]P5{m?EhfN0E`50`zӡon8c8?Bh?eI<y\[w՗XoǏƤ>+c>MqV,Q:̂2j:g0QiWo (,CkYEEV"kG>BqlkXM]C 7U0L暩Y2/0<5E,Nz[? DqDQ\Z~wD!b"(m@uGXd u|^<k t@>flXS P(HyEK/D:"X%ɿQ9YPbR.: tI.š(c_-*[t2b|=< Yʫr֬%oU{#$̊Ik1RlSAZIlIbE鹿d,&د_0>1 I1IXW5_6QZ;K?WSQZ+#{`[z%l!7Y Vwyex bRg7vtF@ |ƼgkN~>iLk]ӾL͒ 2aD"Ft,M`"fq;U jq;[ w` Xq0/:ŝy6} ~,.Y|ߍIwoTo??&,"}_hx5Y(tހNgo$2UpO]gx*7830}n340.'gѢ/~xPERUtR3'iP-xiP #+3FDK [#跬 _aR$ e<ϗ)苘{"__N/>`$[-h_os4oxG,`wi㼑[~z{ycj cUtEP`_N-K2(P5[Ӧ7MMmUi/&D@&ewz~:UAݰ$I.a0U` \9 XD- ߈I, DUvug8Ï6V C+EAþwge~,],O7H5 U!Xrmsji`en i Ny>u`a~,K\ B,Mp;2{+.{E:wB}5,G!j@5ӓ:bxS{Q޿DVݾaYPvtsׅPeԾNLiO voeMT)@LՖ7Ķ2oKmhfHe04=9b$YȢAXk+([nG')XOz.Q* /G(a!E>N4.q^`2Kq%D]:b18Tsnɢvג j,1E6$;ղG)iT4qSG/0.%#<27a̯,XFjV_S_Zs֖ަl.mƨ$u r%E,R."{4V%V[Fw)F"UcՖD$bjLg+Js3pP."ovR` V)n2잸 YnJF0CJI2zݯ T64QkTessX΂U2N2w 2Wk@"R]90/`\&+źvR!lmh'L'#f: 6{7<[zOʃ,+&6WP(I{1!0VF$%K&a%j@Yшݽ7Ƕ6/A$*CdW5ԚJ>).SS!hc]:DRbNSvnO5j6D^% Em5[ɱgǾ0<=~>LO22Ȉ24Md4+ILO9ĶkInrdYV4?0TX+ "NLaGGUF֊tX~fY++s* K[kðL3m C;O= =5?pCO㚁o0v\&!֍71@iȔ9-AZ#NLىOݬ dܒu;(Pm"&Q3@/++6xtHvki cKv r3!m \ŒՃ7`)xLJtegWm?f&]/%zOzmʖH̖[OIdІXѓYRrYjJ]}~ߒx;4ѕyA`ëKmE b-iǎn{qxqQYahRZYM}c~[l#qizEVW\ Ȓ|X "wYdyC$,G\G{o3],Y:kDΓٌ㣦|M ;y@:Sh`'.0{u1'̟G.1G'2,`6mC,LTOJ6 vr$tb3k:- Ǻ&D(džG:7yA&46L5$We{{ʲ-RQozRTm7OJ"U_eygo2]+eW #G$DVg( o̗^%&k&LհG+p- [f(l}NLC8hOP0%L}4@d@n(խ@x(q Co뇙{`Ü{⸆vdyI#[t\r؝Cz YysMVd@1l @ft 9L>@؇0pxһq 3mL ,ށw\x@&8p>b8Oner5`fH, PaMqFdQ, >ǑktM"d x8N- mg yU`|sBҠQ?SgMhJ˦x%48(g稨jB5jy$ & `5f?{fڵa_7@'޵ml2 pݽ43QΣA׎Hf#d93$5M9jwq9yvW+I>̒PۓPܚ27BLzj7ϔ܊0OwџYgMS=@%}- /tVݔ=mX, /xݼՁB`MzB|u.if4Nfs.c>ѭy1xU A!nVоD[|` 0x6C2J2M"ϖ{'%(.xA=u8bZ^_ML_^Z;s/S^+E1Kz`&q[@Vx`[IKR ҂ n`6x+@z$H[7%IzJ {qC*yy"<H%%.Tb ̀#WX>ymvq._0]xX<Z3"y P̚E}T}$Kb .@= F`x6u$U gS#-H0" b%q."^#{s#9VS~5_`E\'r^R f@%.|"74= :ĜrRrAp> ޘ_1ŹfU9!s$MfPʬ`sM19|py-`N`t u,E=4S`kuM'&d p_t>@! wo5zT{,Gc9Jwo615 x6䂑JRztvh۝OMg~'Z$S7,mtX(i^]jW3 NSI9(>rOZK^R5ONN^?)t:̒ML\!3DOlʜALQ~W!bJ ~rlTi|\h"*'PblAnk?/͈f@ʻ֊h[| `q1P~25zѨj76] .d {X$W*E$q( e4ea.F~TeQ8yf#oĹ8Ir}}iⳔQeXfiDDY,ei&,̒\qLro8ͫ|S$fZ:C ~cɗE{f0cj2' #6FO~gYY~/ሀ!]N~U D?bX hfs!|hɧeu'i>"5![mJtG7g2aQa>^ٶ|~qvX^v",g!zB|jb3I$K ̆na8 )aT1G!"?L4Oɱx#OQ:vȇGF\,26H򱺻 `%+]Ay2/hd!MyiVgW{ttcĸ*B 6:Fn} MۻU7M{GZo<)t(tǰUEaJ~5jĝ]>x$A$;z{ IE]~&/oWc.lcD6 fקN E Xȷӛd,ǫ ^ưFjq(QQy#7r4xVΰ֒0"gR͍*7s?["'4^iq2"ʻ b3"n¡n͠b~́=8`uŪn>{5ݦ_]*U)ݨ,F_X9'f}Ong7֏q:z}~c,(謹8gkYgah^  +/fvh!?ZyQ^0DPKzG򾜗_{uĚ@=M'ϫtpc+WK}sSQ  /`'݀-#ZAE-AQ U e[ʶ'=PvOBuB #ߓP:z=Pz[* [*'TWsyyEE\ Cף)SwdB&[`p'k&y8߻EQe'4@^Xͯ h}@[;* Y 6 sv_.W*ļʮOD9?7#qwHJ̔-sp>oڡf'j*RB?17xJ܋y`gjp=Q 3p~Aku~Oh>/ljP:N5.:W5Wj!)Z^cQ#+`W+-L,VYqF}pK+yRb:YȒ,3UzmŲT0U>ͭ{GҸ6gsJ\u=GiLڇv, 6hbYb*_mѱAϝdǹ Ց{?(0AdVUӎ `iS3kmL0PZ<|x%+2"eR<.E[T)\b  f Եpk}Α-޶G[!_>oA۠&r}?@v.OH.su~#5M) ]2 VQ8F8  ټvoUzH .곫kPh}傜  VHyY/\%ypf^nw9>uVt;Si]prvTB%Q~X۴Xt 4qDsVACd'V @wQ!m"j\uSM=BD6 㖉\3Vǯ}s;n3´֖!TiC3cG!7Uh=?X7 1t)Gc%A\-/ҘN}h;)lS^32F5Ch;3C=32s*; Q_~A-X)adܭ֧4d/I8) {s{VԚT,X)YY*I8n?ՕrΪ@gJK`T,X q2`YEgڌQOA͙d0 f:GXeqȜg4ijZצp]t=:vѶ&!Nc`>y-a4_ߙRƋMJ\Ş)D$ wR篶ͬc]s]n6\Xrlqh(V8T:Ƀ:8ZjϢN˲ʊceϺ`: 2vEU4kKzB)-")g>a+P;R.ylZlGYe5ʱ:ۮ; a[I/D7!b] A { zt>&7; 8,NI'HLJUJbyȨxԥ)BQ\o ~; ,X` 2Ĉ" ߷oD]: D'Ez)=>w ?5#bo{ׯ7zeX[ &bU>+Ǐn:Bs6dpo7kb7-x n+^L7|@׫~sM>6=*Eὃ>pFr.h{ /}䣸gḼKۏ/v9vJ2ciȃ Gu\b})aTjtU7 g3C\4̻E@~ C Cw-؅ǾOaUzmZt:(5?Z.bRAv.bQA_eFT1Ǩ۳ftĈ;m 臋 z(vF|LLiԪb2XZ5^K0ؑ6bwb)p1HDY;)2obd*dr5_C0X]=~s7Ӵm̎ ӲLR8dF⇐vߢ# x / a#ӝE0Ҙ-QJʮ.1F%_nHpTcaB"KH5ueYgc. 0$1l nҲn0vK<^5,Z%dJ5Q-&W!F {ody{@2֌=Rf/Mx.2/Ky9HbybޚY 01]P$pRyKj:eED=6كM-Z/BSPIIVc)K!{!x--& Fҙy|Ipu`+8 dWsulQa"3XgjЊ9m,e 6GVٖFvBCqt)q T.hճ8B+y1p., j(#,,<dd3O:L5F[==R(*_[ >iwrlvD&)jL{+X҂i\!(;i|lpÔb'$f)EQ+"*p6䖞$F -XI ^jqVH})#qDH:uY4mwG߳C= z}U)Y)~{<Fcv.Ⱦj]s=?wIA~+=5:H58v Lň\p'+D=TPbנL]^!" 9YeI[ hZ80Z7TDG$ N熯!i_Z6=ET􌕡gZV \{ZF+r5,Z+TPY0L!`\3ZUOFxZ%#t TAL`$Q~I Yhq$5ASQ"$&ÈM~4(FS":pA|a[gd#Q7<Җ~JMo'pڀOvLVD@1 =i3| $8x<0 >VTlcVm%c1ΉH748l ``WDE7zvB@0k5DVX0b0b[<>QdVpnLw|scaxlbE0nOw^88)m<` V:&:=8.nt`%'w4zkKy]3zn)F )0çӖX+ʃ8f0'm˫byw}&qdmVk%^mY(`iEXF ;,b-$S. 013)ߨ-ulb3WcUݘ scnhRƍZ6pZtmphe.xބԠ0I@,˯CR^tB_5fa,tI#({B"h'\.]2,yɫKU 8Jҥ$GFmK+jny;"sZe꼁uZyDyGRYXM-)P QGVR":$z܎47|;$浽$֓`[UX3 J\U]iǖ~&aWq0Ov:jUΌ*'>ŲdݧMOEcd0TW!1! |Y׌P8ƹƈ!w3)ljp$t}fJ0oMI~" )AMD yUYGz)BH[rH@Ta88S}"A-"?c/ 9Q<?xƥ}YT(,;}<@Y |~ ,[$^> I0%I.wc}Ij& at"tQgd8wι{1]Q3]Ŧ.~Ohfl"7ɴ q.aM, 5Igz' .m>>cȷw8 9ҍѧ{F*Y;V"0N'~mVB{̨ F~{IGpW oمnܖ{C`v?bqGH7Cm2`D+b4L8YH:;O;48wZ÷^#Y/1_c1N6LO&XJqk-8^*,b$<\q43PC5xRq36K$qH6P yIxx\oG\~6'K+ 58x)pGF jn5*Y|/?.ci*p!Q?l9͞KFr "ؼw:\P;hQ3,тb~bb}\3йA]Uu7nJJ/4WW4Cw5Q=8&tl%f.5^HpjJDQch,x,REI%I_7AI78Uxz5̓TH19w$\3"R^ǻ% { r^&ͬR2 H Z&88B yP{SbMVɴg4˾O{ u4؍ߖ~Lslc&uxV.7|)"ՉAUDH0eV ;kdSR[?HZ NN^ɧb-HF;U ;T {xl4 tlfL_2mҖ# Sk[w !*I Z1DgBT3.~J\ 7$cSɗ28Vbߞ|Hp<,82ed̂-qeB`8U#VY- ڜ'шDc|`U7SghH 6&TTwk $y-% A /Ϻ]1tKઝ3 / "4x\[+x"d5ՌM~[ZƳ]|<-6;_cׅzU7|ǃ.pUFѳ^ג#')im29-߉0n ?|xEػ֟Fl,%K(\t#d?Pu=+f aWS n3)t|:;ΫNUɆd&(( [ EqάށJq0D r J`|ӣan$Af:X@۸3Dr+z ii"UK nYg}xt So %4Y z Һ^ ~{&zndhc7[$kmK3Ѝk? Tw bFΫ\y?2ˣ`p~?Q&&q6Cp2qu(RK(:a#= a 胟LE !\6A$ZdbőO ^N2Q<6e~<fa8*H;6SryK&Z aWƿȧ+=;=Iڀ0iVZ4 Qy~j+"w|T7g&Y%"k[NANDu pxAォsh8ǔϨ;P#i=&>߁IvaL!g7~&;Oў׽O.n["948-aW ׊[tdgoJt={s?V_@BgfQi |~/ )neS{p#{/" Ű`~W_}.h|Lpf*. G,'z)emߌ'dr53@!n]GQV}_JmlǏ.{cзӑA ǨMU59BqHxx7݇hx᭧Mܑ{VH..~\ލFQ5H7owa&b kiJz3 \%Z].5yT7u{ E8,Q]7Sh5JnA,-_L;omwZSw6F޸,փfa4,-հ!f2C^/neS\T*s3Y5 UnVǎbsR"1+@Wd$[6uPRssəfa!ѧ5չpKp@طcɵ5yoȬוɮ{;eL"֬Fޝct5ǽ@*^iAUT]+,X0qDWE(TgF34^^J+1GG+-?~ݧ6TcGҨ"0"ON:Yi^^A70a̓QnR-91B}S{8Υӻ WnJ-u4gI _A::zBܿRLj7>d*%Z5nO'KD[R?qzHJhEB+P-z8罊q\61Y3c b{_rO ˗:|  ]>2$JeZ2J*qxT` Fsmrf=BIMxjWN[54e6i ]?Z6{mP â/-ņ͔7Q@6b Xh Wg Dh4P$ IC]n4pZ$ ӛvhל4Lzނqm6 ?fi-XV"$4kr%@bZZ6ȱI}v]AYt:#4^ Ep1F(7xvz[W j8Pwhգ@UlBk Yߛ I0a g;,JJ2֗+^!Ccq,JKMLT.,;KxŢ,ZNhv^y{ǀ =]Dp qVJSiF:wO՘]43 >GD* 3ꐳHs`+pʶSE'U fz$Xn/>*jUYOEB~~v|T!^"_v3v88Uoׄrn^kcϷzy R(B ԨMFi{6pm^81zg"c$viU^5!R$wH<Dc<:}zs(+^AbSUCpJ2 a{Mr- MV'p! q֖Ԏ$;K6Aoq]U6n F-%G$i}Ԃh [ l`6 ,8>HR {gSvE~Yd`|QQҨM; wZ\nxzu7}_2 QiCOo#mo@s@}XAsUH=f|gFfO>M3Ϡ*zH2v;O?JWZ8LvR~X ej{|4@A֜5F!lX `&Sz~^RGSYo tqxΈw9"od[7yVȐb%hBԳf5DtJ "w?}߭r,Dxe!qgpXmse(Bh9#PA)4\Kӈ/ɓWޔg<^PU")Q퀇]ޛN η(7PSu2_RPB媳c>Ƀ S e~o:|gXX` p67RYP? wг45'- Á`ܢ0<۞0 jI'g|قcg. iJmP& 'DtN; lma-kΎ.UrmQISg4#a V (k` YVJj* K%4R$lɕM AR<РL-LdՄ[ 7.{W+!\eAMҢP޳yh6%i͓Xdľ$t!QcpAŘhS>m2+Lc\"k;4/ZV0VP+`kCDtg%Gbݴh\>X2'꣍0J _(Fof<ި%dMbGe;6X?5(9{ {acܟXO9DD?Q =}j3S﮲#ͨb -h)^h>cbanl]59:#KICwVB7xoymJmWBRL\]9kqx*%5YcrQA`pH3eLƈӆxZj:wZy7{}B[MFK Nƻ<![D)mKaK&=E Y' q*_tL8ca(.7kj@(Τ.Z̬I_Sɞ ٸ DKW#C.rksAiDs$WKF55&N"(&5kE3Pj ̥Υw,k%#QmG\W{>uM&Js[# Ix5P2N*E!w2) QSP?PV'i~Q?7@1:Rii9F+r|3=īnp* c%!. 1l:eca NX`{(8']# iRV)mNnH1ObN+-+4c ^Hax\ҫΥB#-I$-ṳs06T,40Z0驢0_s!t 9K_@Has}k+H|(r x+ (99t $§qᣱVRArV+̜ebS+@R;}r琐TIJ^ ^A$P`9$7NrPZkr0G0EN NO+|Gcp%jQ\?o?{6d`"~;H2v2A&oVew3"춻%nفNZ/"SdpwIB4o# Pb)7j)X5LɌiF(է0X#!@TD+L06ppFV軑} :ck9X:͓ T1H];Q.4*D6IU(\$@ap*&e s5:L3#b&8'l~а5I=)ܖ-ZGS+_kO 0k;j\N > z*hBvL8uphD@64:QZQZ@K7:ȊPs>yiL'9 1Hn>k-83%i:0&̐lyvhT9mFhKcグZsky *RJ6B(DJ1ӻǮ֪*FTUHSf"sZ]j)OVDXcje“8N1aUqr:UD>Zjk8C-RXs G̨TcHJ 2Q˜mXUBNUa{&F14aƓE@_ (*HҰZE V#[FXqrN(Ibj,>É5$DGL7*s$ℸ&'E\Ohǻ>jLG 't5"Ven!:0ڋkZ r|㶣%-TG'Gxq>!]lO'W]l 'W0dBVC_o%<@ê5m1&2e/5@b$P[ֿL̎ X-}pn=UFq(X#ܥ. RoQM`X)3Ԃ)n=OUM, 5J|*D2vJ+\j|BJ3t2Js\-v|hљFU>f]$@fIE[ }rS @HuJqIւ-5h" em:0n?j`vp[NiE?މH` Xw)0QZrUn8._lp۲D &/gwٝ_$ȧf|Ǡ|v??-NGzUi$|fh-*uALQ3ׄڝ!p11na sgiUDHp=.7P V&z0 㚀:qeTI5'g2Q!#Q> b;<:KW.̟b~.  =k7E޶i~𮨘KǗM'WSfČ$1(m?`O, ,}wk?yځrgyk5+0%>O;o6.0G$~QoQ6̦9XCm^C~ԁo9}ѿ?ݘ;'KhҦtٚI7Zi|y7@Tנ:ykpR) jASd ν JνŕF7YHt IW>%EKdzׂI6>:ծX&\2ծcBK!7 swKr(S6QrgJ7Ϋ)I0]71 •1Y򛭚&UJ`ToӚ\oȇl5έVW!= .$%N5\t۬nu8խ!zd=9փKG+үo)QJm%s5*^eA|[z~͊c"|eLPs>^lq6"1/ks ~pidbk[fOVʙ`M 0weY7w[kUwnF^YY?ƻQT]n5mkL j`^pQCEԽ.ܷ?RV.Z0Ն$Gܽ0;bz2KFA'{^wRrjV~ζZ+d >P f9qV`v濚p۶uВ@ck8!B#oU@Э ·V< lL)ԥE38pUٙ Dĩ Xӓq,7LN:-v+µVR#>HS.m{Fp*{s]J,{mx+;kɝl|? n!Pb!{}u7^vrw*_׸ ya\뇾؞ /LgΤ8dY}NgYvIީ^[j"M5Dϙ({G%#uPt ~%VǓL/_`9㼗ҭo`iNѓ R~)f2ү.av`2*$eq?FK ~)˓FE~Ĵ HFƗx$C `ρN[t6ο_fˠr5@e)&S&| cw5]4 BV ?iR'*kk3]W/hEx)a2v%;=~ѫa(yҨHN̝e3*L/T= 5Љzbc¯o^ L[ D-%f:؀]xN=cK^-TĮ(gwSl'Ԁr8=pR]ڄ(SZ@~[Aj['.F@u}jf} N}c}R`ܤQ>Ob9zaߘEdi]L"*qr4Pox)gc3 mI ލ }bs 2-_\?ήmt#Y||jݡI*]h\րǸt@[c+}AaMs{oMm8mBpy/:AM1(O>lGvrpel@2:92I;O7u6«`*1ƌKU|d`2|-%g1 'aZ+3b'Dxޗ &Wo/*qp5q<"l8M\t2zeDܯ 7$ʾ׸WG-fHA` 3}.4HHW&raAZkEw& )EA?]!~Y64g+N0.f2s^ܒ s;o:GsA1x8!<jƻ9{ó.hFţn4yp *- `=bL'8”X2PKX)bՄHc eDU8UueXY':E!.QȌ6I &Mb˜l^y}, /keslwhEweZҐH4XNHZD}5.k \Rk{DWalx?Ap`~pe}p1y{󢔻/Эm{pe >ށ%ᬜ^ M>y?6qY \_]_>7Ǟ7kҌY/YXJb3t2.+ ew7p%p_m) khhK$FvXphتTv`.aۃ gvk5=Gѵs$7ýq8h7 /奆z ۹הf#1.ͰϙϙϙϙG1Vx<Ƴ|)Bߙ\Kq+x'יIPejwDk(k8'*3L4p0"T=n9)gP!Bf #JL(0"e5O\B:X}gTI¤MQ0E4։N͉0g6)I &&gӐG4LXV 1Û!r;{z?KמِY\L)ʏCţ0Ol?b]1@E식[D[D[D[DaA1?ԝ2LmH.v*lloVv6(+Yw*JSOn929L$7,"$JѐHRELIA"וtI>$9I9DŽRnbMPy,N)Kc<&(J8%Q$"D4=5%v[ŹWH6=a9(#S={73Y(Gpkv915rݙT{w'#)5c\+S'K! M3[%tD1Tk%i~fjOh}ZmMZy0JoϮ[WKʳu,mab˵81H23./Q1OMl } M}y"܆^B4 8+w"a8?`8_S(ߗ=') 9R:dQLB% q,^OfVʏ? !QZ'͡Vm^Q)Ħf:SW+VyB00…84)S2:!??*_(uǛֵ`K-"ӯhUlBbExB|:#BE(̅!ak'J]񦵭 CC<rS1/nd=PS檇)ǘ gmd,V~y/PТoE!OqǗĶLfrH&s"5mӑ*|Agթc*fQ8Lf2^w61vr5:ra#Q9Q)kAi,M:ܠqb9G8,fNY.)=RU+=ߙn~ßymr&!(p#1qRqb9G8,dq|,eH8Lɐ rCXjxrU̲Wqcʧ$o`Yk|"߯QIGc.aE QYc0|,%kavi _cÉ"pO-u A?fԉ$4QI0ɻ3}m'g3`drMI N[X M%m**vm4:#pYrl,Q hT$0O2~$q$>bv~̎r!S,j\7͵-ǫ1# |,t';M"f!!JyAd7mw)#7w|F>,G\ETv:]v_kÛSC<x!eQD seQyeb6K:Gr\뇦!iedheqPZI=Y{iHV#0@r qcY>Eo i"n6f)4FaL#]XNn\z_/fi{x bT;Pm G!<+S&݁FhXK7gu~jTۣ8ukR5RVyIJD\g$Qiu=_zݫk X{i#>ԧٛ@ޓhVsH9 q 쿧/}"L|4w϶cPW_^1뿓m&(|z?[{MjYpd6j~ju.~9qK_0}O6J*oňՉ\M|3$ܧ!p?YIK`6lp23ƽ]Ge+YFHUgt٬f''&CÕ DΎ %G,;(L|BrݖRc{$Uot7pl9Nx8~'yT$R҅ y"(r'Y`0d+ Kw'AcVqzG+꒓@/eCB#+)9DrWX߯[AeTQǧuWnZi'VyOQM ]& ICeNbV6$.Q4 8IΚXcpVrʑG(U8}1G4GՔdo@-&kbqiHi׀Xo@嫙/f@M4'hInՒK3ʱ_xU9ry] Hk~yMf[n|{0R&5+&,|[iEhT\ocj#bt8au 8[1SǕ 0; GOv`AE@ j"Yk?sT2CD*- =F,cfƄa)8%\O̗wQQFͱn>;;8@zd(k;1Hr-\V>)z>qNfr'XC10HCWDz\JٵרXm8 J$ f&4 Xڤi:@QFЪQ]#tuM~@)qV\d.\:yZ.5 Emoكj %r{@F+my& FR)KmݱbXk]kƭyJ ?^fڀr>_ɡ}e6Kߤ^;4* JDj.-Pz)m2ofx 4*ne2jJ5W ;8f^@z =a{;( 4mJAXzy|q;wk f]ߧCeGE?8A7·]@O^UXE!S,Op[.8W(,73VmWo\RQsm18`X<*`NCfqْd};N ך 'cY7(UB ,O7SX)8b{P|||* PL 4*ejCàynkKz쌽i␊NrFSNyǁMIu]&,Qp1Kw(u85-cXߜz&l\?BmK*.AC13~EDM} f 3`vYL3t/cdK|&2eۘ S,^ntQyN;#SJ64p.jGG(m0oRn$KGBۭ5>x#ycV =C;)>Xy[}LVK턏qOV\mYE]_[dz&4s/)|[cYb`:D'mùk@CC寒3ĂGfPh7N^`3W 4*~Eي wjY.BGcE 2k_ 4:UW%(,Uu7J3j4K>yXmq3(*\CIBG~r5A:cFL#piXdlВApW>t1ݏh%vH+Q9|'7 $d@i|,ҭC CDk:ϲ21KyבƹU8^cDz\9 }ZUz=!b OoԪ 'W_S{;M RZVsVE<5j9ꙬvCԨ3|z§bTʙAr}ow@T"gy]ax4NWSψyʙf6|5$ {R}(_~䈙SQP?U5! c*33N/t9cPHCHIUUA}oNŸ́Go}G3%Q"$Av&N2L:M(\~]" ^}Uq7s0eb:6Y1m%b~Pcks')6 pQ1ܶ(33 @˽!pso(,wtƹHmk֔抰Le`" BDQqqp5WN,M^ȩv\pB;ňBL8R\Ҿk+ _?x3UCZs)s5](nMU;(/egwq7 er 4*$A!JIhTvP:0x+Ί8j&uK -RK9Դ]LYWkVlF^>z]Fŵ2 ]_*{AMܳ_,}U=mӝߥﳢYJD*fql!`!UTg% 5V{ОiC\; ":SУEYݷ(lp|1R$}m}@@EVOXq;a,z0asvV/M5~0&ܸdbCiڤ ?݆>pw/X.Q/|[07|9`"/r3\N-~dOq#r@YQ Q%^Kx>ZxةwJb,.)UQZw-flQ%jCo N+i`QktmSkG"҅ y"ܷsZ WV6fQ☽tG$RͧQŢW| y*N'TpI[ -m`9M@Wr8i:@Iu ^useL*"Ҽp'-qֲt5[D-gw~oo3EDŽF<EQfP%~ tF #Q~- $O<%jx3fCBnt` j&PD4{'0p3nFiQLT @\M4 ۋGa;@w"k&|>ÆrEo"/%(GbuE$z{`%OlQrfcW.5"!$-̮ccn&G(b pFȏy^}±}`vd $eon!|6lCw<́[~KhOv`m%{9}q/= ) :g| TѹJƽf_ ~; :mDF-I:tN:?Xl'?Aaў~)j\ n \?PRk 3")LNuBy#˕ϿSOVAle1@eT8l$_!"g>؇zlyEIԭݙ,%%FXYD{4QȸC˂1/ҝ/^JNj?jS;pQ.D9*񤖩w'p{khmB[w1@1 s]ĕBx4iݮG+dݣ|:ӕ QJȳn*{".8!"cR:&RU\H\D/Jky_CCzh+(kYI| `{9 reXՐdbȊp%stA&QUUĥmG" VUh=}ƯEX Vtӂ,wgǼg2'zB Ag+,䚮)qM|q+ÏFMy;K^K(tuPoORrS1UB΀r8]ҿԸSF*ix ;==7.zyH=|JU?)qho`NH|NH`(GmQU:Gh5M&ɳ<Io /s NLؔJX c%IT:!@VJ V,3k;L0TfCfj̝UkܠAQڍhPs9_!A Nx:6BwY~EWS^jSПmOlI'Am~|5^;`^ocj& yUD9nM0y 2z!}me%_:$$fC:F OHLH@yY^n`IcJH@~?5td*,|n^%ϡ,{Ȣ d$ր#;m^ vA$wGgL ţR0+61P\,Lr'HW٨.j zmל=8kz.ncj:ܰn/(pV*ms<#=]*j^lRTHQ ø>q} ^pj `Es5粃uɓ#MzE9BǨBR^. X9ޗF F TojxO:Fuȭ| JGa5Pkx_R3Sp5 >li%0!1I=D%, i˳> 7`i"/#Fb2sXze81jdYI%`zƱ 5PL#t6L>қqR,զ ΫI AZx_2zO DZwMޗFB7+Ssg,?&HD~Ab!< j X nD?=cT,X\ ;$JhcJ,DP"T"ma=qut}F#pk](*TޘNc!lG4&As-N-^-;%`6HഘVEfַ׾tw*pJeUZ+nGl3BUHeh4x# B.R Y/GbQp (陽~/kV1{% IQНtY]7nY]2\K1. 4&#C`Jyn,< B2L1]+.]>:̻ům"3`szrŵN˂q QyVWkx𼼍Q)645z]UMGo"Ï[hV(ŊírQVUr̭hebߝ;UG#RCAUu!x-t ̢djdFPraleUIYuC7a\C9tychJ0gs.]6w|s9d|[%wV׋mHZ Vj( ٘lϙ8+.HO70 U~gC+u! ^(Ҹd۳#X{f)8V4b־̺Tn|i/>dech 9PxLЈҞһR w"@Y%%#9(8k0~!-:4tbsIFT_ԐNL*Wd4&4ahԍWqwvQ^$QrAF#So1gGy6>Qt ~w~Onc ۀ)^0sJ,4czcp'-(" rq.U 2M$!CL%38=WQ{kh-@WWP(^ dp>/gq wxخ(5}JIHEsk" A$b$Rh8/tu<z/` 6Ƹ9/mQtIU呹D$FJ|T2Vs}+;%/4a`0)xO' .Pf( FTa/l<qPEk4UxOsZld9KNREPF'3(meٙݔ׿~'!=,x9x0 ;ق@rM{\rqsvAOsvqTl8!A٣uDұb'U+{崔'䓦mIMMjaZ GqϛA= mȰS-73dgkG 0gLRtOkxЁ|5wbt@%N P ȧMޓKf5(%Ml.  Mr(a}kt0t@Ů9V>ͫD<iuv)>u?!`t=Ȧ__Ҫ-鑠z;S ;䬣WNzAX:೹xl0 v.)K]"ӂ`V@AȚiϋf'ٯؤńk s斤׸ΕWd]Rt%qJy}cdjv+s3~_%k.SH9I5)gXˎ!o*NWՕ!qޕסJ bsRl}aIO_zёayS:J&veXqz9n:ZGq\u0@Kq-@+m:ަkbA)*ʩyÕ`%L$mKxD3TwT<1,q4C?vV4P K*CwءBҏy^F[SUȹsNp94,o vc}p J/M7 4`]R~Rd&7lJTZ|vptaB;P|APCsys|]j! GǰIf{#g;ILSlraN B-8%$.f@482Sȑm}J34 F?I"l.>r[Aa6%^[tB@u&C8VKzNQ7*Ӈy"CGݲRg+A +$tZ"i;TR6EMR!1h 3%eУ ~2CEavsoC.PwO'P |3Ih0a%z{zYv z[p KrӞV]:}Ը&_RF%1!Be&H$iT u!f.10hmYxLKUD җ`e:|UV%%^=+2WzXt` ۇ_ ١{(y aSuxy acP%bd BD*@B'ǰJߞ8C$gZp KEY>κd-8&aKiyvuycX[;l0~Mcf:oȶ8>?+ߡ5(ǰġ7p9}%3(/vJKL4RQHqBȊҁK1?wx]qr;<Xt2|‰BBcC'kK ~˱y^V:C5Mtl[f3T{I R8>@2, r,>0J otN#p>c"Q5RЫZi<iM cFu,%1JpH m AC8hWcv[G/A2<&"̾ &ٴ5%8nF\MIfAbU]u>8GK]ݾSq4JU”xZ<3-#қm+r -5s)sk:^<29IsYM3̋"޿)2a"YJKT/GoHtgoq#tLCUb>R2 6}ġб~2a%-TG\jczG G~WԜ۾zWa$1zcД|*z@FuKʆw(7U* Cqe` ? Q^: ^j* Ӽw̖OA,!jeQǛ==SEکB)p&ccFEVjXie3)BHOMiX{@ J|VLxQYTچDe^O.')Ez6&1M.$(q*J mG(./=tQ:L=Mlhhr'hg8㍴o=؁#.qHM)ARgz ߗR>gaBuɼ`A絧$*'[qc(Ry0UxzDvޞ7=51- QQV](pb>VOЕw^pkJor+eDy#.q8TYްeU%H;e(Ƹ}( 表^cFi 誨\8G}L>x9 #uKf$q=YP)0 C!S=XtNbi0}!Q bxx&iubD-ʁ}릵z\~}ī_7.(Qh^.8gjڮO/v: ca畻t8<Ucu!K>DNLPU (=vH"dqA@[H*37-:߯.z~UAEшaFLȊU 9(]~8G@#-Ҝ M=" RQ`/W hIeDA8t{i5cresJ&W!u kJ AuMY"hI`nA]#- k C/̀ߧ׏KBG$`Fo?konaEW?M~?-ՆwdE)= k=7O?Y|CЛQw|0_}a>?G6>M<{0'rfmg**MjdWZl/t;Y)㞧 .vS\Q; i]* ω$4N R8x>ПcӇnL3x<}3cM- kWyrnI0[V U' [^AV ūRĈt7hS_oYq%pxǸk7+OU3}c _ .~G^d4+..}չȗ([ykmD46py{ ?%@񵚃U:zc\c4qKdc@[m(5j=>*=i>h|g> BuB jwԲ+ܭ 6wm{$ҽJ0=DYw w1w,SMhOiU{恀;yP4c#Vl̴1qT0ZQvx>mkG`$cH1 Sc2I7Pֽ_H e!ߞ `1Roc cDJ%%I>HI`Q 3Y1ϐ**34/J$\E!EOϳٟ`TIv!SCjH¸(Hz_.~X?w1z(1\&tTaF2t †҄ Hz ͦه.Х.gkXktK`.lc ]jâ(@! rB3c(\Hx>'eZ3 tmpLLAcGP?EyfX(rC+\#O(-\w򪯏^VC|p3\Voc Hhj˻D9Vmv}8V ߍԆsZ=~<-XGGGQ .}Lֺ<[,*NnОEr)\g1ax"8aURVPfȑXT%rsa]{V#wҁCR&"@-e$| w|܋7[ |^Be8e.K;άDw0MɨH \H?9ksi)W7+!fCRۙL0=fy .Hd/,l =ƶovk[f}2 ųF- YΫu6]WgKZtibv*3 ~&ax%HaH!Μ.x.jznuXxAQ\L+\̦W G`CG25U,c Q8$u3r1r{INpkqco-!(|kIV ;fਵ!Okma 8e;GgGtϿnt1<{ 7N L%6cE!B_̐+V0H~!y yH IMu_a]om][[H/|:rHsn#! IXIJ`V"M>GpW`fDK NmwYC Mm1^U1?uj VU?yr3z~?VԤZLɴTOr=/M뉛lO[grXOC1YIp6hRpH45*:Z{r㦌j^Mм޾d?7ּ ~1_?L \pCtY5FvvɂPܻ?T׽p2QKL8ubqgs`Yj?FՃ-?b =[ k.g_ɽ"[[/fnwzjp#,pM:]7Ц߭\q'Ct; riݏwVg5g Gx lTRAL9֪30^w 5MJ;/\xT[OnSJ):P|N~?&D28kw*}tw7;n .Γx Gy߬S_N.o>6]-ZIwiҹ򋷑?a &:zq;cH0ZSuJPP?`97/W?u6km7ؓAkq fz[b;i#c;!ݢ%'mRncV5 \Hl0f~%yubXK6% )y7('8dFRT~ĕttv{[GfԚ> 8@ԁcK(ۓ)sz1źwAE"f͟1>Ї~{(]+ ҜpbO[-ޱT_fԋyJ-?ʮ< u:~38̭l\9^+SbFˠ('g-riJ8ޑPŵ P~D= OC I1~qllmoe @B^? ¤D38Q̀5-^g0x(]o",h,5 ɳ Q.Q\VPn:;EYbq*IDaZ $:Pr:}e@b#AUm je-䍪Naw5#)(vYkIP %fv qH AoaJ5oi1-&嗢=kρ \LS5]e:pq`{$J#CN!;]~q\0vujbѯHt (u5NΦ颯K\-:E?ayK)qD!CNs࢓^ O\>}4:qgd4#\:::'xQljҁ#2q㄀TN_)#DQ#עA5+ZpYa8bGrFzQMfaPgDH3}S8Ry9/m2krnqD&v9y/Qux?H]gdWgᢞ.Jz#Q!&Atfx-Zw8"]U;Q0d^x 9O m,\\D K+N%\~~|p֥2QZvt(!&:em8*f߲z=2ۃCtxxу,'NrqUJűWLJȅ5(I0{"-^6 /f) 島遲=b2 N.RjLL]6N):Qj 8@}^}?~._T;8a]3k1:щbW xxOĦCsUth3f;D' c$5׏O9+( vj{ݣ& n S e&ˉќ¬-đ|4=K,3j+Gdqn^IV:pD&6yQaN$g 97 7,4XbTZ~iCŢ?Iq]nCL`o7ؙ\M "K${+vѲ%qldU_*pv-sλwduU2Y"f70 vLyg4yZIyv̄ nϡ#)T@7<$[m}t̜ hnM_G1\ui#vu2"UVXV ֗UHS .o^1s2G !&5b$sg7s)KkjȔ ~ՎyZx2M9x9jgPB1䝃5BA.:m4-4a`2Н`!9pP>e֤fKVxxF!ǪP1&v0rGpP'4=@j%0sFg'f1g.) q9aJWRD"lT"vh4tHr"ģXԃdQ3ͯ@p:㘂xF0WH  /b9]tMe>alc~[IBȵ>/G1/*9F:,H:]HREU (F>ewS 0`xG6r3;p:O-s\f!Α/b'Xeėraɘ[|E BN]1szD%zܝ2D3[J4+.``_'̎|ZU(p34) [Gix\łDP~A|aԉ| }禼 _\EG;/ϋp >0E-u gv:Dj:|g1Y̷{vEt8tPy pqKMQIf6ښ1n1Z c0a -b ^y[׈,հN+pű s r;,__ǍV/ڻ|{t2.i{l_<*bʳX]-p! CXQ$!VWL>C-J[og%v~tH꫑e,?Q@A-Ϩ Dw 1`<؎c;!$QSLAn|e5㔀|y\e2,{h <Y˟[Ώg/j= 4'w/O3?#X|oϦuK2 Gߦ>Lc(uC?O:an6WK-:kz›14hG/znpח| 6:WE?k4o /ͬڽ[o/˃tfo t5&0(~nd)d8ޯfu2_1۽l/[&?+hi0ChHͶj+9Ak~ ELyO1R!eZRb6L,A%֞GJ[SpV_%=QϮ? y0U @h8ݼx|5mwXTʀ`GVøTG+Hok=j<w꛹qxqܞX5 7-!gATQn͚z5nen<)DO)3& [2cǕ.*eSq"+^IEIqI1Z벢)*0ڱ즠=Nr$zvn<aovh/$+T v8kںcPgQXFj]> H cBXa*p-,A 3Sٳ 'CCSk6HѡdBsn }5]UDy &R0cbZ`m{g8wGvۤw\17>5?Ofa_ځ #HL1`Qv$0K;0Qƈ;aG\:pK%OnY?L'.j0HW}|竃`U-ۆ=`, [> q3ڜ lCJV?DSB)!6dwn;nkaIR;a!moÍm|X0't'Ƙq@8k}3uPpdf52r|iq)Vt9Kz8 ? x.z=;f!9)롽y=H~@/8[tX}r:+͛Ԩwve^-},`Z]<GϹQt@w1j2fuf([~T3 kv(ڑ u)ZdNNbn0fyXQVKК+ ?aݫ埈>J!DoaذsAfߜ6ņ"9}xR[OnHpd!Ia)iS [_q \%χ0/^}S4eVL7LA}D:]1l℣!IU4X˺1zpz(z3%M5ڍI :Dog{r|&HJv?Y]cr=4joU;gu0'VOyČY(W1"qQz'DZv.!Y'Oy/''( ^զ'nNZKr'ZomN=n}yt9pSzHlV֦ZRfVm-yNB[2+{:;s$e77'_K;C|tY} nР9|}]\Zaܜڎ!r[ήP>Mfu*ָEe}U{i6rcjq%x'ZێsQ[Nx~X#B)MdB6 on7p3 ȟ_ T_!Ǐ kf[։  1w-| DNm+YOi"[e0IDW67&ܟmf 0t M/2s{.~:Z|$1j.ۣ~7-"nLx/lysȣNyusodLG3wD^nEһQ!a,D NƺTߎ!oU9;fКRIWW p艦<>s}>y}>WxYPZADb4X `UA`j@f3\+,|=.]|}>uwGMYw]|}|4kv]gzj'`L9,:/eY>yq ()ϋY-d|Vx+F#j@ <#`[PO'j }B>vP* fވd*_Zw .8 ?7_`b^ MAz4 CufɎn!y>{‹EG/1mD<EA9$X'+Ch!>,ěV\ć!9qZYgT#~ 9DnP\0&!uHW<eZA<"wJK*īʀЂRC+"x#~ Y׭y'LKa.: 3a<5 nSZB2ݝ9+LYn),L+$bDzvee V(}zb?bX03T D<T #q.U>]Vn)}$E˅!~-Q#bgR_/fڮbJ6 Ȇ7exX2kY^BȼB9!/Z nSK6)v4ʑ2B m{& AЮi'J0FTh ̑Һ(8m^%8BZ-EeYt\ft]0!{mANeEV5}#L0’ l9Eifǃ1Tm(Vl'OɥmDŊjm'ᤲYNZLDhEȭ)#4!8לx$S,^h2%ut Lj ţ9oke?>>\TY^3P3-4[vw~ "ў_%. cJ`rI +p%>%5zpoFSQU^Yr[}]qUtL,!~ &#ǿO|_u^!<6VKQ,)*Wկ [SoD`{¶>oB(ˢq@VW9#ZF/ _#)#0tB5 o615*8rn-Zw,mIV`: |qgq # ukLXDEIhU@wN 6iz#Գd}MiD秌Ђx͓\D SrxJʘFhBk\H[hGJh/D1"SFhA|&^`HhKfSBލ?a[޼R%L5-yi]I!ߔZt !14JT2fEB&RSFh@#ě{Ny{~:elsd 7elwZMڻ՟+m| 25(AV(aCʾdNK,)M\G^w-b==ZϗKU/Fsi${vwVO<ˋD'2}N>+6;zWBzrͷVD[= |6;-J:_j /n{ yw?q堹E2y|ΜЦj^LE*dlX9^nz +!Dk2f4e&|TcddPo xDGJ=OL<]kʫGzYB:V9e+ZG"W9k-IJkT@B%#)#4!ު. ;mJ .L.Y'O񠛧fp.0]%VBI2Ft~-G.Y&R͊23a *#ȧЂxMo4&ȑ_ٰ@e"GB'HmMnZs-߆g9t˷trW z|L[Awr+%f% t9DՖp7ه#KBZ/} cT)rEHGYWM݆h+Ue?cX W \ۿKâ Pe?(PeEiL"R5bǒRPf6b6Kq/ﺿ[1஍T ~8~7\ Wڦ7_ }fk_L߼;$Οzz{k>mǻW96pI^3 VG TPIX4wa dSK1U$zOAt$rPH*e{U^WU9zU^WU9j*G*zU]Ghys6{f9=gl͞s6{f9(dZ9.fg];KE %43dE[KPe'c7v#w5 7Q_-9]f8ӯwpvHLr$7 CH +o.ͬpMm߼{w!soIuxJqvρXW97*h0p pN/)Jan[^ڞb;]Ǵ0mtt4ev@yWBZTSR;2vRiE_'fN(RX}N  ]hGF삡hU-"w-juC zO?`t`4  7:|.PItI®-4u׶ kAzqk#Uz͚k˾=`5uU+jЪV'{3ל*l Vj iFBQfX)<^lP{fw+})O? %=ړW[ŗ[>7;X}q0 ۸;';ik0.3׏J'HkL^S5Aywʚ{7/O)xAnBzAz J%GAmA{L7s/lb{P' pygƟ kh K`^N> P{c/C%E/XPDXX(CIpvSڻ=mHz hg QaL֖w"L:]@(*Yp\hTΑZRGm\w%nҏ6E7I7q0]u+ӳ~09KI{Zm:PV:PC:PCv90{9' |\W]C!mky:m2WPhbwo5Tx&?g?Ov;ᕕ薱s;i9 ={?<F>92ۚkvM!֣_0a quov8(Rk f}cܻƬ'Yڗ+um7:&!zPUKH=Qk/K_͗^@ьE.-,r X+,1,H Bњe$6R ϛoe3-UY9|Лh8ѾtF[RZ.0m;p'xXޱw,zǢޱw,#X@<^fE9X.;)4[ X L% {łi߲~ן93 QIJlAPy9O`tN%. `0uaS O"Ut%cDF/}n@)*Y<[_al6:َ\<>_ViV;1`=;?漬۽v_.҅l{U^T.0Ƥ\pt19FF d3G,XSW&ʻDo/Iw,կ%7P f 37(훇Jp&UuNֺZjҠlɗdPIkP6$G681ɐi!h*#Osږg9 \FMch\J4̬,h{O 1T.B 0%:C5$DIv1fәǸJ_L/?39nPʥMc<wQqGCWh`òܼpFՐSօs[t Iߜ[͵7)TBreQs("欍1#f[`nAz,-jЭ 6yWNQ1Jc˭:FdOªxELsq(eB؇RV|ne" S.ӖuɁ/z NW47G)׵=a΅." TGEQ$`"d.{嶍eܺf_\9NU|%.լc.2IٖIUYhgz6cIk}`+=:$R;-#5HѢۮ4BI+ceXHI'W[0lmK"/m>> +]geMӷ;)hAߊ'/&wnMM`E^]T/v)\6x0On`0!T6Ҳgb|LiH4ZNL,o} s.ZK?g=^w';G+ݳY6kZ%0\uÌGȥ b'/]+'r EAM?qu׷n~z*}_~}[LͿon߾~ {؁Iq0@qlL7O557*Vej^窧^e^UKd-/wn[s]×SwRͰngw57&/ٟgOn:DkPaBEy@Yt:.QrL{:D9­ws%"FB`d@* cv\l]Ѵ!=r9Hgr̓(JsGX 88m<njT>.b k֞r23ɷF?&XQE/}}ȵЬI!Cgμ1X4nm)nI&ӝ(N!kJM~dKR_(TD5%J'k&B/6zbaX]Ƹ0zymn7G~gh?6վn6MR-RV>Gd̙D;zI п͓nngFd8݇0r}i=hƇhd&Z6 }B,HsU(NeR=Ðp=PiRFQy9uApc&(gp# ǥchKN?.Iɸ"wyttz 5BFGy4Ƙcdt߫-cgmI gZr9Afx[TۨQ[C6y=yZ~[Q@1k|%_&~&Z&UkR&UkR&UkR&UkR&!&}RkGIնOj'ZKˋW/^v%Qs!QƸv2ᬵڪ{m`via50,cCY"厅-^N,),=Hz6M}^͞Z'LϵN+~=9 .C0!B*Caj3jX$"﵌FMFSvvXwJs˩ Y-$CRTD ר]QQ8@ҒP]¦ybHt\.|f=a[]^kD@55Ug{FJ||~y}h-R+)&[)Q[M*Ffd j0DZx[Q-wU9Dim+WrXAZlXZky10BB20Pg($^2#YobA% 90g Hc%A˖ZNk9혜vvH#]c#4 \DUjTp!nVËtxۓ+1:Zcx~F ] ^rÜQf YcI"|!t{m$OKֱK8cB)7j<[֑aYqGkZ H8H#i$L\' O,]EO+"dc Fk{39W:oWw"CO#"y톞sZ9T# ;eEXLI2fl,c‚pD [î|3hNLKh Qay2rL(,mG- )M тYN{Sngb1Cي>K1g_HVٸ`gtOƳ-ֽ?d]\4>ZNT41BD3u[Bd47tc-)l1Qg(4PGy^:C!1FXcOOh;3SUg&>t} w6NE4ٽIӨ .&;< ogvvV{xxϠwv{,ݽfh>KȪ>y-[*nzHŮǛ=do*4_)%=xʑV4g˟Zoal&}t× /+'Z|NkB{ /? :!t|ԋƃfaZ*Qlc^VE#B]0-v(`߃]v5Lq1 c_/clb!OK4ܝ5藬9n/HHetTj]sT*Avw;? ؓW%Y9#0F#c:Ecc\Ѵ5ucoG`m\4nOI?R1q"p8Y )[}s_ )_4V:oWBN#O/_R<!r: ׬\^4wE# .;IrlKT,+Gs\ >>>ֳi.+ 7NcbvfOF_ TvZk'&''zOCà;;6%۸I k4/buƏT/*i=2L FI^^51gcD..\]9Q xeQ/[)XYIm$**(m(^Qɨ>m*GMpF HM$JRV,N,n\)|9b0 \DOrYKBDOo4EcR5x YǛq'kc_2N(mP-UwW,|):\>ť`3J>FFvԋ'ZO_/Q6[RY?%06Ȳ&ĚcyK T媗-6FM]86<^Buv^%WCAUJL9Jt6 X).g+7^0v DЂKu܇, |^R ^~>ŎQI\eeDw>k.( M HC(2 "v!),3 ~91+?5V!ʅNT'b?6 h8a{$$UՈ\K'!)TÜaB"CD@hLZP[X$raK׷#/G+k#*pU;Bgh&_?>r ("C$y>e92ä3ɎB CR}H/% z"A_wٓq<_6#n"F  \T0E]RLݕCUll?X0݊P& -,IKsp#lHo$x7ߝ_1A|;&g]qt\xq=_/a!e7rނ)go[:.;<;3Ȯ|YYNzV Aߊ'/&wnMM`”́/Au{KboU@j7A͝7?2fZZL)m~4|i-0a.`.&8g=^w';G+ݳY6kZe;X8U3Is #~TX_ΰ(!`PY *&(׽r_޾?Woo1Q7}%a&eR vyi]S|bUuzUUu.dMb!r5g Ju<|y9?u.= 1|D .XhMͻKz;~'6y?S7"50z\"D d -C <,4i:?{Ϣ6 ~?{{Iv7L0`Mi"KQN߯(H0lVw׻}hlk;kn7wMɫ|Hgq[(`Pc*0ZX`c)3jetX'*KI5 qL,^cΜ廳eq۲$⎧.y({EF`_pqQ~ /yI۷~z<4_y?'˫wj±+sqLe|(S oI^~b;-&i^涃]jc}e;7J^]>%Zmk?:#-ttY/7\w/5 f/'u/[e2aMWfǙ}:Ȥ6Am:thD|垐{MW'o Jʯ5ݎMb›|^1anLsv<6PQeoh=6Q3!el3cs&>T2nmx'uU>p]6sQ21FxݨXrJ+ voy\|<:3!>SgtI;Bz~[2vހ~i:^5ͧ>O/-N [8cL:BZ&wx=A=ʴV}vs.?,Y^.YA.fka圉B : J8M4hz}x Airg\,6Ι 4US yr\\ΔȎ:m 53ձ-(mwc5)zVDӫ:P@}Q=m yH 'a]n}!سA1k*A7 Y:Q yݘ8Iv sv/yg+ܺvPYt/+G4UПYF{%yK@߮Ehh#3 ͻt .ƾ_ƣd}^]l85[> [Ms12H)%W#zx9A]'QӇPq9sCmo;iu{d, omgM8,'b%ΒEɧIqe#]|.ZJP}糾AOE&vp3G( 8CV.Ej[خ->gc0\ Z~:WO Cӻ< e4IG A%D{i%wahm8D=Kg;L"Z2{ ߂m ߗ؇rٶ~M4G-#O;.N_c©k|4B~J8J5=#*(}-=h޾>k6M4S$ڎ3NUZݨm5D=ZJrv+Do;g+@oCz_s6NU4 ԕكzՅ?: rne4 D}QZw0K}x֜j}Dޟ~e~{_{;j=W3jӃamHkPNuW/Cl\v|xf-vsQFG3h$p$k0D"9i;jߵu! uLZ*ҥ.#J^ݚ)aw7 S\6Ƙo,:9^_^ˎ o~~_va{?L܍J-cR(*q 3][;Y0ܙl͚5%meяD]]WŒWNmmMR^z >8c: M07$}atC>Ar~?/Bı~|ATQ0f}:@[FKH5DD`wTT\DƸĕ9TiXXeQ,9];ݵ<7ʏi8?:%7p180*qX)EV2 #TD#%&P`J}8;d=)@U!-?t.GI`t7LW fm5nWˌ߲Qg,5xeu| .Ҧ|1i+"]hm?Lx=JL0^Sb5l0ܻL skGWqr; qAͮF{}AG ,}n/0\6yq8B/hNich&341`b } Щ6x.ei Rz[\h%@e C-YZ-^b8@@2v>=GQ< A2WȻ1gF0f6Jtۅl%0dμMq!?oL qoO,COw7OyNXtl 5?BTs:$[pdyl_5طY"X@1$ U#{#3ނݲ'x_Fn9; _2Dz1J}?],ײ?\BN9ysLrlHy.PHʘ@G, q9iFXp4ztv~;Mٯ#`(kKx4%7Ksr7ÍݍO/7v-3#_lG5 9}+x'344~_rؿmw79ƫջtí_5_ HQٯ6WhyGvM7ν[lLρ o{ﯽ?սN}飤*{me'GA[Ub4vE}?3%C3q'͐JgM'*A\!F܍MsLýw?溶O}gC_p4_,0׆0FAV@w74ϣSqY>,>oΊ垊! &nGcm%04(5^srRb bD2HRcC(Sv?&]* l͜p4fB$FJ{F+ hŸ@.O] Gg>X3 TaJ\ۘi+<%ۈ ~m;om |-;_bꯋż[/ hsoҩL57ʜfMckuπ&ͅ_UaTZXJ[>`t.HΊRCN dHx~a[_k-rѢ+-rѢr(GS" 1 F"CR+ 8/Thpl VR0<)|fiҚԎ^DbNϔҫN(:C wWnz_]w;be)%"L *Se&!sk8!E(ivtw~]? u$9%>Mkە"[M> gл'?I\0 qigm( XQ\6I.vo9h 1)``M1o飞v0HHD&Ir'QIz)VY6\ˀ+h YҧV.sDW5lF===c!jٿO)!r;DZ4#P!+ym Y%deKIXc-M@uȠjt8&Att3I&0Z;Q60Z?-g糕`˼T-mwC]8o3 Eb]*KD[A#%jԊB:ݦzx녵D tC%GtzW% ZUj) KuV;}@C)]=WC \]4=UO|TSS/''X=y=m+el*|Vr񱒦_燧ޫR^(Y1 r:Dt"/t5ƣ]}},=' Z͹ıG|sRu՛`W;O(ìOr? ]ڼO ~ˏC+RFӔai{t蕢]٣4qete ͻT^^2iݓl}x{zw52r*%/!w]ĻY5`_ZA tfKS8V+@_8t:["!Kvzu&}o__ Ņ˳MG1$ gBB!scY)S#ILU q  }Fʄ62Ҋ$r!SyHl#@Yb-B$L)@I% rLZBJN欳S#ǓQ$DJrܐSO¥(_<^xLk Wfr%bM08FNrGbB* OтqTBB0Rˆ~:BÓx$֢3>p ݅2= I8ˋ#yRG2tYz>OM 7, zâ~`~"D`i j=XLb&A* zF0aap) =Jgb+HkPΰ "H1`+HIXA, 4.ARllHX8IDDکGNB&^wiTs/YX3˒1z -t]aʼ%k_Z&>lIh\ j dp v¼SvNK^.MWS{]ݬ'&[ap,]D GHVzʾ.+E)L!i-'5G|tr:& x285K6{.?Lj[7on?,iia٫# ~aX0ֶH,C 3jؔmGp>㘽dXSxluȶQ[֪Z٠g<H>RBcssp[ʢ7HRNㄤ&Y4W?^$uᕤ*댩 h\JHadԜRBOsG Ն$rBDMP'h"I:M$8Pꉀ`,PGß4U43"WeƠ|U^&\-dy,蛟./3`sxάfyW瀙 ۑq&^R@r p)|s|s41^dAon+e8ݵhB2kaL3ڶ(W?p亅/5.3 c<cHօC[ČUL1[O5-/_L-$ŵmz֕5l büϘdGmg\i]j<⤜rtʭIo-0>Ļ0~J6s:cĽaFZvR;6p+p{~d8,i)D<7iP~OPlf#qW٠EblbFaU6]+a9-8,I#@)e=)DxwmEd muaOT, "@&)Z x#jgi*|N(- N+vHNh;Slİ,Xl#gM7w(CTm)O$1Vd ٽl>̋(/ZƃIRe0xDE\3pA`2{=e1PIJ*vb wLd.b >IțU8r:,93$q)("2 ))p2ILvI .f58ʹ\JLks$o :kCX!5R HEPqwǚҸP4l Ӕ4AT50Pٹ O̬gN!&17ljxW]s& G(SQB9aGVSlh6<;J#;h-%'Y9XFH.$fRML.WR/ɍax_?/\B>rt6U^;.sW+v0/՚x~v~zi@&`|4; ;=&5>nW]]V55}j]֧y,3c\ssM}z$ْ3pGr{5+ E`:S7^(_ BZZYW3k;YeĘƓZGa(K>.=jsx69FDgꢓ]vUCU_ge##-d4,>LCy +WeEjtcx&*įˆv`?{﷯_޼7`uPwM"{p{}5Ms#viZvwivoUWvhai[k]o,T;Ottu=U\|dCjo,^TUX Z힡G}t]A*%f>-Qks0I pVe.I*7{!ң8,tޠߟs)ӣx\8e\*)sAeyiq=Lc̘kqecg&t_;rWҵNKp Ӝa2(`{&x=iJY 50D8x@Q4KVLJ>Vy|qIRXu%hYN D<#>KIKYDH̰pP]KzIT$(t$Xj ,Qk"$veT hβ+{}GpFv8|%Oǟ; |׶y_c=~3}&[#$ ZhmVJi̖h,Xr9 dpۋ9.ūMQa-[eXPZ% ~DY͙ D\%)vB}:d+]wŁ: ArMtnʏ p$,t%fJ_۷H7zuw1?mYA jx H8}J"*rxJ}V8:KιhY^:KtBH=Fk ʒI$^2qc( Dh,PG3I 6*b"AlTS Z9۱׍_-Rii鑋VFtbi>vagal5tD3քX<:MFiZR[n` BQ}%ɜ$ $/d&KyFu2tDGڜ]Fy:k@-rFv~Y> KJhiM \d % GgiFd:g*EKZywuDz2zy4UˑXkyѤXX}>Mj S}ֿAu{}ik꛹uq<<8+]hO7Rqw V]v5n-<[0T7Iky\c[,uY\Rv++j lx",=+IO94^IjoSwp>{XFƅk/?KU`VtAnr,h~w<)iNw˭u۫uST `^!WO,XxdLX&j^^7L,S;Ծao'pQerw"{}"C7_mӃ{NZAWw")ۿX i*ݿe`f&5[[jƮ?snO->0ImEŎک×/3q7mXN6i d8r;^μ/TZ1օ>}틾V34Ƿ] (_vNx{Z8am ;+&Yϔ+ټU)i\IJm6TǃRq*,WDyчd ,27+; gG_oaӳU: 5Oypi Aˣr|2mwՋ+NM5Eoż.'U K7Ƹ_yA2OP{S..)_H.ջj]fEA}ReŧL }CSoVEfj# (9DȣR}#9+n;pn S65SS v\LIj [6*Z9asN1s᢭u s i,O{Cҵ2)_xQ~ +^4|#qWxt6^ޏ`'P(]L}ȕGߟ~ZZWy:;l(vIx)2u4}N.fuEr8k[bFFCC)_̂`ɿJ|12 j}Xjq$\Q o BQS_0<5s9ˈ#U^4@/iZpgI3>&>!l*_A^7!ܻ;R^TeVVO;?Ց[5b V4PYhQ˭ X]N.KRKa):?Owצ^r])ل.Q7!hb6P(]J3o=r= O /oeQbmQMԍdbƿs[f] C4*5sJ~6Y8ݼussݮ5X[9w<[ڮTq1OTZuɔh"%̺cX:y2~\Sx<9i'Lf/QVWudYl?2mrLlIUG'id*or%2@evEr{e{r5j?izyOԕU8 Xv4YE+^nNrJx%uDrA_IF ޳};/ID5R8'.Lj}d!rV2SP+ 掚qtRRynf#JOjuJ.%p(3xCmJ} KHLye,wVz=Sbh 2r#n]!zƕβ",HdDKdk!`-vwk0c3c1 bl,$w(*A`fa7)vtF 3BT jCf4;I`"0ǨS C ty ZCDC$w1gsY;\61 {CV =H2jDZbŞJM?ȅebqR(ծ{hIT{aJK1xvI)$ TP,PoJIH) F6˖z F{m!Qkԇ}hD-\{)ObXUjAv`$vqhvh%\x)V"\$Tך:`B#D %za-4La/=heyF e4e a[# &` ւ`2.e`;S70 K p§(RyP1%p 4dX *.b!5dUą$6+uԂ#J<hd$H /*X Eh#ṵѥ%ND XJ'Z YFNx2p*X "2f7!ڠ` Il7K4_Dϓ@B]K PF ˤ#P n io cH$H0,d@җ6 J 2[g X,UR)C A " -L #^"&VU _ awmYevTd6L$|1zZKBJk+=դ(JbS%A [շo{NU[r L#+%po]MEwF%R92rܠ`5pFY,"=W/: TER+ٕw`mssJxlLbgVUQc%Gji"j8(c e0>j2a+"Ǜ ~\uXgLӄa 7H6[v݋~2mƬ*$c QIbYjÉs!u@es7֝tt6!t[ۖ* QA&'z|eM=OUpئ:/`:pĥ8(}tmV28Ut)\]h)7VSc2(3Hv9-ȗXP0F4DL iY!Q`rQR0-ac;7"A)ЭLx nEf06]ƂQ,TG''biVp6SRe%)8 dxon{~!O‚.&jHmR mZ#f@ |tm^?bTnDl \*^ FYuDhAʀvw %tMކFL3yz5|< hwX@zT̨D#+Nu:PDڤ pd."ΨA`&Ybm4 K4К5''ta,o3,4LkTRp0U U-eV hߨtf&= H/ԐN%Ta- \n F`ⰱAs`bҕKB/.&iY\ӘI(KD0umMW-d0L&Sns`*]ӿ4mF[Sn I<5jhƸ7G=:6G _6A x XRcF6JtclvrA,B2!S : # T", wB)Wvz M҈ +Ip xrp 94R&xykeNV(h(cx)U,JG52⓪``u ~ aQ6GV"5R&0Rk+kԬT gF x+e<(0SLGjJҸh µ5#Ϗ;jyѦUC[xYk}(ѫ`2 G`93#0²B3 _XD[biDF:|%q<{Q|V;ͨ>rI"VŪR盉`ebBv"Jv^E`$1D%N 踜F LBF7Vp]ꊄީED8Ø1Հ]mj۠:˭q6;@/Yt2U,rRvhYh_7~>Jʑj;"[XJ]jG/ģ&hNSRIhk+ׄj^C"nF>nMI뀮z?`< H 8X'xXn'Jm9-lt~rTG8Egy J;%`@KN:^Rdzx 18}'[_h!wF~w df'MqÉb eؾhZD5yr~󻉧9= ]z_V\MQ ǐKpꝽ?9OwP/GuۇOt\!Jv~^s0o Utrv}|' 6`y0fϙN*J3!RaL#\\\Va>Dˇ>[^ -#s>0"~rVV v#C,QM(ű{l2JVh]T޹XӞ\}wKKd-?*}Jwe1v?6-FGڢlglv_a˓ftz]h\F:Y,l`?ddBIUg[U(_7-&ƈ;ʲ_ؐ;F2mb0’Edk1A+sa7&hǎ)fi4h=H{# zz{[ M}U ˲;7 }z;-. ^].^VNEWgQU*JEW*]TtRU*JEW*]TtRU*JEW*]TtRU*JEW*]TtRU*JEW*]TtRU*JEW*]lU\-_SU,_OF ի) k%{EWaTt[, iEW/Ť,p0^:*fJt9فF.(:&XT*nQim4d+A$[I]l mxvk2r'E[e?7̏Zom#͜uף#+&Fn^Q?aGyrzz{sn>|uUGc'^= y$ @$$& 4IIMh@$$& 4IIMh@$$& 4IIMh@$$& 4IIMh@$$& 4IIMh@$$& 4I ͍[7[nNV av_S)yu׫?~5#ּ"6(Jvֽxvғb[Tl2FWlMQ[ mwa;g%T NK' gD&aXB ˾9-ڦB.[+p2\ہm58Xp7pv'Q^רumc JDQݕaD|13_t-~ixp4ws˫{n F8|/pu/>H<5̌0B!8թb>PJ{+9"tVE^QxnEV^>/s2>M!@ ӎ(Φ6j0̉hb3$bEbY[ݨr yr4sn=Z毠,d6{i:d]^[zh{{%';@** ߽i&zmKWȝ-fc9mu'MV?>67\ WwvuzwFf+G7qe+޽]{ˋoo|5~7lG{q1V~MNexl-'2ڵn6׼xKw7~ɡ7k|I~>~L Uth}7&nJѹU5ArL7:%:% |ѱ+9mYVvgt;S:(y{B RAkirzw6yWu:,cWwC.{xrU'g345mXA)1"BCB:{o\M :-wʊfU1Q >EDFb͜;N12냖64K&Cj2i<_|}7; {N7p0Zΐ|M[5kU\5b 3Rl}s$Tm%5mwjѺn[5~n="x LrӆF% V܃./qV#/rj5s-F!j!lmonl&8&JM_*5e qg+ϼ4V`EHwSlθdOQ)m4FH=V/4\0%E'.V˾ݘ(M:;Ox{=*?)Ɔ;F l:Lg}B^l|6=:='Ǔ'`zCwxVAd/u2Ye6OWk3|͖+7X+\\ƾN:Hw{?}~>;;>9*՟t5,J[ Y0?^J&j/׬/>8`y91*“zv_( 9YQҚZ+lM 0dJU!h@S5Q31/q {\qeqb%<| ߵ^Uܵt-0=&WTIR~{JJBu).J:Ù[:~>~F2K.MjS,VYmg g{ȱW)Ytټ"d:&AӃv'0xj˒[e'qY]ld] L2G \GX&wLfR:4:#g2q2~&G_[G_4 5_Gn +6: g?xWT{d^(D\ŽV\Q*=`;Ty;az'Exճ)PGwo&+[0"J?s[&۷H gD;Qs4y𼾅c2,p")gy| 2UD2VzW_AxRYoi.`i0"JU?UuJAFҦ:ڏ]>wh?vFΚg?k\ow> Dܞv9wDz _v>7WUpF'R t* 9H\MD v*w]l;#g5ig=>N -aW}_{Y ~k'gSrAΕ-3b9r"6J![Pp3ŃetGg5 jp*yi*Zĕ rL-aT@:٭E]mYm{- g>Y*-#J-q7aZ|p9bw׽8svJ[Gdo7SikD18*yzWDr-4qgKV!TbrHR1F 6<6gU7PzPwڨ)mjD3A0ɵ Ts/] MeY=cF@1gr NVFj>\2w4d7@/})mȆ|4Kht7 >͇Y6kXRoC#m6F-҉ko]V'39mZEAXNHG00j!3wݻMLD*+Lzf-54|o=gex: spk}΄&A[bgx+~:7+ؒ<774rn3$LH$F9F 1%/T])XA<ʐ_+aYXi#._2Ff%Ĉ<% TٞTS*X#ˈV10f '݆KrO4%-aJ.jj|00d{+a[zBl?%F AyB&0j|(N ՑSǔ2eHJ8pp )}K脢k'TQZD."Ɉ(VSͰI<p3YF#נ %MrLAhr߽>>(95[EtJrk8;fǠGÖhe;#i%'S#"L` A U G/M@e|,@U*ɥZNk-z0ol%~3<Wn[xݟ4]s{;e ˗.lvaQ}g#mV? b/٭?< ^ L af vMMċAY0=.MwsPi=qm>seYM7{7[$})=,2樽]_ z3sZbuMvRɾz\r|Y=g=cxYWGX=S)DS˖^W{uv*]EԽTjf&!>i.d\.E]ؾr$Bཤo%W[l/ߴ..ùM~+Zl4% S.<2 =^BQN+h{fbEm7f,EI LRJɰQݲ@u>+:6~={L9EMTQv\/Qe;ݚJ3:хu[_fq+gԞ>Ĭ?qv8 6'yä aR R4cQaad $F<]o~jt7Tf>q7ѡٴj1LiºIΆ#~A½0EN93XZeO FF F I#GNSṀId +9^h V="TǠuFJqTQ&"Ew9lӵZ"֬%RHDcF[g6>=i҉>Ǵ~DHaw+y'yE{0?* DS]t`(+(RJRJqFr&,DG }m-})GB},ZV3%fGɈe,jiP`aͣ#(y货^Hlovb$3AMVْ"{vmo9O'2a[We|$SBo!z3:g&-&a3rfj]jQIgk()(ؤQ2J9G?wiQ캃j  id  |m'x"O!%M!@啲v7z}>.겪_J fC5o ~sy5}d>˪DI3ʒzΠ4M2a1_M 'Ym\; p+pd@a%OLdԹv qSp”}a=-Lq^و}*--q+E#"!|<$?6I~٩7,C0!B*Caj3jX:Ak1hFs+%co3rDبd38y>:LRȋ U.I|kt"Y/}&V+]/.>`Ժ+s\JZu?*Q_䷭NHDJLd+߼ċ3KW^EncB9#:p3y=?`lҷE?(.0^s0.IPDM2;cI}`+=:$R;1}d7T/llϿFC Hgj0UHJL/2+HB $kc`ָ(1Cqa\ۋ)U4oS%ynGeQ>g d)Ʉb@ N!^/RP;4E!x~:5=  !R\ 2p'e x$0*ea,>OWg&oe,\g$*7+%ыGGr,$BV$CV+ ٸ:%ǠSYeut*`Ol0^gnpq /toҟ^ooO^t:y_N>} Z/߭$Iu#0ڀ[C=] 5CSw9zWu9qo7A9խ%Q9pȿ.]7$O:tњѕg l~^m]T}*0as! lmq74n}xEz4oM 퍙g(TQ)bt*$ $E |0 qXEEM(<ȼdcs4Qxo)88'O5#E5v:9+ă7?{xp=Z*f--L:e/aW9L%(Gw>ͿN~a;i 3 &TP9 \`D$yű5 V `-M1.(%FQ4!J4%ASJL V,LtܠNB7w{^[e˩# ݋Ό2Çc>T)(ȖҾؔI.U΃PW<Q9_䃹d|üD;e lD,F=T InV٧TI)/6\byU ]wC|FZM= L,Z /< .%唳!a3 LF^{ÃFF!VkY/H,RgMIUŬqhT.8`5g\YK=XDpI UXPF*j2s^cHJOIsm4PƩN#ߝjӘG"hQAPFGT@T KIj@Cc L"zzؖ25=f+ YU+嘢*Wy#S ?+[iWp~5]bRc\aL )&9 ]S{hahWR*/8aZ$u[@1F: lp6HF7g)Oc.*5Xz:^{=cezӺme%nہյba2q]<ZёБT#pa?4ԍPQՊ/h㺸c0dp0@%.,2vAI_[ڤ^]Uu`mSP;v(#V#m= ZG Hah՞:#,RFXp4V {!0/톘.k,/ ЕkaЛug>O(q-~:}婑W_ nA΋$H{GvGs3Ym#hjsH;kI _e[y&ZL狅ئCznF*k ^moޑO~̕Ȟ6( u Q^8!c/Ф KRGoEFtiB9e2WN2:q?JqiTgӗ>5 mo> "Ν|*}o)牔uu]M˶>zO؍n'uM{LA#4K⥇J}^86*̜-z$fo^=+Ȳw.[:7͵̄$Ղ+sK(f_CefY1bO1"묡-zuqdJq`XI:D9Sː{uv+]'ԫ\jװƮ\Ȏm6.t}&DM)Y)l#^b kx2_1c{vKkp}O S@{ojtF Lkkf[R0Q㲉iCrWnҜNBtN2} ^v<(o EFaaR*]^Eը\(U3Jj[g䭧 &reSRâ׷zIٛAkk7-iwfh$Q??V ujQwl !:i%'S#"S"`8q:+C*jPکeJF(իJH1rK%bdF/cL#յeLq5J9-ldlk iǶf[ݩ}Ւ],wh5{ۛ?]MrW>oƓw`0T@O$Y4*Z$R:"liM`'LԅE.IA803  D9(#A1]hسˬcYޑiQzHLdu4(&)%v.teA kZִiԴ7a{P5["4ZU=/*Y9{H{3S΂,O)P bNF]դ z'* B /lPS iU'oKK0ɼU678S~'*'.]yS2ri*Zĕ rF+X;L (ZИn˚ldvpkh)PQ\˄^@'|`)iMKmQ1De#x;Cek)n?9ftt$tA齡#pa?4.P -Ǹ.%X*V P9 ˭."iG82yyym$*8ʝ7\ Ø8P=t:FY^1Tk%iʦ#5CvtB ^ߪ`"f:_,?>C#A=xL3yFz<ӝg"Lw3yT@*ňtGeLwQV;t<#} -ͱ`?8WN?8g3p?8 e3p?8gl2b:u,fiY2p?]h/jts?IA% ra9- -AƒE!t΅%9t%#rJFH:1 |=/r푊39|!`@}⛽c@eEXBI f)lĬ`‚MtD 3S/~;iwܱٝd2,gJ8$ jiPL&32<BJלo_Rg~P™0 n8 Iz{ELFtɇOpg3CC wnsv~ĕ?i7iɞ;3azQJ&]z680>sA)Q]V|rHS?8֗.k*8Cm Ͷz+mCmfuT*\0U_g> ՜5a4Wn^02z=ڥԳea4~4 vyԫş~оƒHĿs8 UxLF]X`$Р)<',UD a2|G>|ϓI?KXBZj?7=OJg \Zܯ/sdF MpvRY0HxWTMi0pW_ǟ-6i*lDހîn{<],S+\M`݋  E q@ݕ3Ffocݷ%|pOnԉyp?t_/f.+yNWPNGC*Y1‭܁RApo0CJ0yhumZ7'w3lOsWWXAV}ȰBxo'l>j`LWƋdr3k9B>J9sh4T{Ϣʿؙ6zrTms7@pɠ6gq%)TD2'Tbɨ40SI%hD0AP4@2~SIZ},TBaPO$<))+?nTR d&Չ,?6 x8a{$$UV#rye.CM +5i\I°TMQr@)a$eEk<+RNJUE}9pI"di"(7< Vz.t4%Zz=H7#+zC;? : ,z?ܳ7}_MI~h.0 ˝!QKCq颤 {/̸)L9A|)gJ_"Ko)L_ c@B#L9GWRP;7qy><  ք!CRť*O!7K9;FQL_CPv3Ax,g.nSޔ%_u{%ׇ ]Yɑ+^>Ѓ,; xlZ!O5-6"ٺYŻYd5Z53tGDFDF|!$Bf\{+-r B͆6 ᬸIbj~JU7]giL;TO~-L>x=$F)LɌ[ۏjn킣Zqyg SMXm]7kY,o}J3f1~tG>*Yhmmu:Ⱥ^k*|tg<:}2W$5Au@ABiTjۯw\Oo뗯~}~?_~_ 8u~1:W$I^`=kko5UI<'ޤ_s5>d-沯4hnv Do__%oӂOt=XHA;~ l:J%UY TX=?ܭ/Eb7E/:5/jQo݊ŔxQp\ Fѩ 0ZD@G!Ճ EBruFJG8Le^Qs{I?$W*ڧ)XXK6q*aY{bgwgS/מةde%vssy=[A5zUt&͐^Z뵟? 1M!rFA*`S-Z^Xh&p `NkipAy,A6r Q‐pR4Tޗ;UM2${&CD{:SDc0N6Dd38)&ݝdFv5 dpDki%yӃl e/;|D>~+˩# n"Wc|Z@QKiQG%UZ@ԝ" BLF'&\h6&YM^el~$`P{ ~&a{nr1Z3~8̈́/-dbk|>A]>rWD SΊFkB[ aT"U ≷x'.HԔ1тFQ4`pHW FNH5'R[kG* HEU{sk,Il>Gbn8iR=+;y-Xhm 2:ڀ-hD`䉠*:d2sp6rsZ= qu!˳g_-&46Eo]oWDRf6[TBί^6[^wtFќwlo: ,v9cAjNtjrF7]wYJ=x OpbĈc< h4jla/ūW@yHTIJ|xW^+kTe#? ^.AUkT0y8@5^˵ ۟&t,¦ A? &hZ(UZFy$asx_Í2/yNN}„xqO{J#:z@j1LQT$.0$5I8;) <* ?s~]`F bXؔ̑@2첑EÇ;XGpY';.+_瞞S|:_\*7~O8EHbDZ H'bge\EM AJFdNE8="Gm7 ,3AH喌.7k)'vƮfKuڥ&5zذMnO~Baz^ek sfD-03IĶV0/2%.K&[q\ʁȅL`30@ L&Gf9AQN'2G *Hcrpec-=! #$$:pFBb%#11)DXҠE0IgXrXͅ@i-vHNh;SRJkJ~m):+ؼQmib0 rHsq1X\!.!Q+c9$*5n~@,)`Кk>D X„`s.`9Vw!O{ɱ$Q -9+R 3 bBMfPJD{rqP0vY4a# ?Y@SJx->c}WwtG\wK8 []ͷ73?Id$ԧ~g %- Hp īNg0ɥﮦ-}6?_57u9*AQw(rLj-K~<*0.zM\=Zڂ- m|,ddG= =jF}rժjVڪemvҵtO”öմNiOF/מkO䪋 s'*1i}?P,9ی1^1{vgW.fhl;BVs-sw Qr.R10MA>TBY VB3 uT:a]hϝGpˡ9-GMba%gRu>gW{94yeM棔lr_u6H1Q.qC鬨rV(Fto_vn+(_Bԟhd|y]nz (Sca6޴+{[vc~mϝ_:VfxAK5 H^:"!h m I%K)%, SU=d,?F4]\Zap?衼_'TtWX;q?OGC*eeqc@~ELm=3{&8`0,|6ck|{w=_x3L.,#p {r5OjzRy [I^ލ7 V|a>ɠz sZD)Z0la%Y[tWr2eD#4v)$-&IQa4z%XQ(DFc-VM?JqTQ( Ȣ`;X+Ж"V"D44ת:'Pjqvx{rS6V`o..oG{K@V9C1E4dЃI*[vAQVP((U28y,D\tUum#PtVbXYΔpI%#dAXPтA=tS |/Dh % DjL%lTtl}O '喨m=-+y"(vCGe +՗v%Taw?`/c>li*q&EMY͒guU=#XkW!4^&qEMn9``v[VRyyKE,!]Xn5xÈA ,0AU*EǸ.>4mLn~k%QH̳',(=iPr2Qճ.Y3IuGzx6B{ {&`hu6KǪ`0.r=7Tamޤ1duƨcĄ 1 O߶7iD#4.Ie\Y{t{4׵slJܺy3dE[k6}us 3[οr?e_w{-Wmo&g4I2ʹ^+g%a\撳kǻY n"SWԧ!=i3Ryֹ穔S.szlgX N(44 :wgLeݶ&TBQKGN O鞁Cj;ggC&2 Sw٠HHK,TK}Fkh{I'OH5\-stӪs831s hy'a Z^xs~HJq}{ӧAcn^bI[&E8pʨKx-[s:髛_뛅ww>|\]iA' ׄ67#c|>p߇`G LtV$?B[ aT"U~EQ0.0 %ơAV!QT)/02K*Ljd<)V2""&ZH0<)c*޶{6rCk\ d}}7~p` x6nx s_!~Sr"[?m`T,(#Vm9igDbn8iR=+;y-Xhm 2:ڀ-hD`䉠*:d2sp6rsZ= q2c9vZ,6EeJey˒Q4tp ˩’ȥR{:xrFcä3^A.GT]HE^/ZT;xԫm,C+:p[6 tzB)#}t E@گc?cS{y o8wYɍA 5h%/Z3֡$x"Lpa)Y}$\.%xI8F.SRk VR5H~6rfl|]$JJiXu>fpmo n|Y=&x"B* g~!F9% EWJbgmhyi?~ |NH:E^c!Zr3KRW,ѲE˹Ό AHa"TWT|B7635g=GHFǤ"9̰4N+!H&6WY< K%Bn=q+lb)(X1Xx1lA04: $q HHF:MMV"j1HS!4Gan4T (t)bCDR;;BjJ]=UfߡsP.IPDL2 "MBcG"Jυ" y+P+z lC6Sy :*c~pfoz}_Z$KIWbI_c}N1 y|0]ά f ;JWG-]zkf+0/=a:$3ߞ'/s0P& -,IVKSp-lH]!E¢PT8ݝwC v_p]Ի;b~^U7Z})B"`HyGZ<j ȊĦ/fT9L~TΪ)*Kwݟ՝7南7ټrvpHf`}9Պ7 ٥) PLs^3 !h8<!P K5Ҁ$Ffsg%hE6ڰV D$gklH>v7:)>ٰ cPYnxzXngwu ˇw^y~7/__`.{vq%H_`Uw#bo ԴaTmmmUm.a޻ŌmMekI@~|{YW۳`uϤ\iSO`Po%T/A*I>7(!vSeN~ZiQ_T|G~C(GNrRT``-"e@H3Z..h:C K~>mMw)88'O5#E5;vFa|s'6vϾ%Xp6-] ;J=;z !iTҢzUwq\*| /tH 3 *TP9 \`D$4?5>tk|N&p R]0'4Tr 9Gф(q@G`:M=’{+3IX&ydc1E4^J D!&"Ŝ$;maklV'^ֆŗŗ<XI./E)yb;NR8{d^(D\5v^T5DD4hG*E#Vގvz5t5Z/hUgfoǾ)6Wc~C-"`n!z˵N# >GtVz8z@9x1)ΰx J5B 2XJK˧Y5SIKjutVۥMk^1Tk%i}.0]2d3RRؐIyPdc,}(#a'Ymw#NMz7ע!`_uwnjzSiYgpc2rrǪ<\ ײQaYV+BS TVmߪKog?E3fW'K ? wHSJŎk#޾z]}^{m][ު1~+(l'Xc=Xqˤ~-Ԟw XߔۚOgן x2nGs,^#mW:t^lָs<1-jEx[r l2DTR ;M%i&CTGAXEKJdW4 ]km;\<v*Ӣtꨛ/Y )K.4QN |._ȁu94m45bR9c6([8Synr졵ku^i5 h7_7Q8vhGZ׎]t3%i7a0,Q#)g rk+C.tlDԣ45]j"iI|,XG(J w`(1Ed aN)wKhTj;%QSXzQ3(vhQs>77:k`Z_ՏW' ʀg2s^cHjKMQ@:a*~Nch2Z`4 et`V@T KIjCe n Rz8Ǧ{O>.w]6{lҾ\cWtcY: `L}O>;@+0? J»?a}v!+nٖ#lݶmDWۘ+PV u64mns -v'=>Ў=ShjtRJ?ROx?=3N:+1y>ZTk^ksX1G0w[t:6⼧ AWe|iw?E2nm~>7{"B* gz!F9% E+Lp]B @ssr׿&LM HwD@4:&eWafXTqXE |V(/=)va>&m,4d,Չ,?f` itHjA2rh jDCGATMMV"j1HS!4Gan4T (t)bdyyGH-.aN@$A 0ɐ2""(7< Vz.t4IwX"VgCηχXӀ*?{\ ?87\-D`ͥH| lUpXqӠD +'(yv2S%xoG(]EcdL&w(9 V` $_:"a:$3ߞ'/s0P& -,IVKEn/!=wQHHXq7y7bu?p]Ի;bi¡ ")~SNjiBFeӥ]o@V\&6u}1SJacztV OQX)o_敳()30zE>]V{KloEjqaCie/MӐi8iчbX(IdG9{xb~"fmX"R\6Lh$M}~^4nFbxg*p{C%k*'+ۙ]]:Wo^//~]y R؁qp]#bo ԴaTmmmUm.a޻ŌmMekI@~|{YWKҷ2Xb3 Vq/)]]>@irW?%@3d! t&pigezMܯ~OP"n+1:R"Rj> TH80"`뢍3q$c; #Q6bXw)88'O5#E5;vBk8S`v9CӨr!̥E+9rYEx:v~[wU{/2t8_/go_hJ ܳ7΍*F*g#qb))yeP'%F r<.a!G< g$ (OsL%o=:XD-f[yTRJH4w {=ؤY"imRuQ]Zo s#Qθ1Ǖp,U/) &r@tHDt" ]&Bg !:i'7:gelZB6.:[;.I̙z- tHI3 `P9#'\`D$4]JrWq+1X'9)"_,A6r Q‐OrW#,9LWq{to_E4'ԦyI\t6xp.SmG+xР@{uB" ;SʭH"agZxRQ;mxk+hWS:L9\Eݟl.^Zv~guA>yn֝ dtwr Ą"tƨ,-hƢ)-ⲋ)-Ҋ)AoXdTl;vv?N.IpqiE9&ީ 22'01dZ⥷$s}u2m_#2?祥<)]ڻSZo{~-}uS%ZZ!:Y2"8˒6BL9cE4`)SG;!8fBu|\ϻ^u>0Ωt5/5^=ަf=d!w}ߓ稼~6+\5Q#mD|t>i;8G=b畻.xy!oJG>6_0.nA"'%=Z4z/cC x?9՚9n͇QǓjSяC?Ɨv:DYeMo޴.}mZx[ 7|C3b6Ejٹly.ȴVVhzR[ZaMTk p4.L.>>hTP/hBF',p:Xtl@:6MUtީ ܲ͠x@V 4_sy1E׽Q Y0tN"2rB'\! DߵHy3/v2db1Jq+`d҂ztGGrE#"#Y*H MTMTy^G 'h1U71 c |4DcY߂6*jB6*Z陎;o|MjPFeB #oLWllTcť1gLz{:sly9@QZ&r }S>#FL@|MAZ0 2Rgڝp}0Ƽr$opy ;_|?s?,\JJfnٔ&| {tI4F3Ҟjq;Eԇ<(i2*> 2yv1D)`XTcNq$uXrjg "P do19XHQo4| 0KKMMgi;-~uOC)N.b2>=܏LU&Mz[$ԕyVKhBoY*s=mkə1fu\ S&[Ai+Q2*3GR,Gӓ F꘭ѥhrSTL ǘHն9 w'J9-2k Ee[(-<93WC[|1֙=rs퇿@x|8_b0ˆ (a \TxTf^,PEm:c$FlrHmjd[-5;0Q Ri0J`24*SL4( Q*!^p>fi*ijhث˦cMǾY(UZ' ٬4QE|@RI1- 8-0PбiMӚR.ZN{U]q NGM]2Y-L0 TPbj^JmJ^n EJ.Zm7m8=YmC zQ=7ȲvRJ a\F-Yd8A#Z p'dKH;yJGg@-R.|WST 7r y['"7o>'mݦO!8vç}nN8G1u?6sWx7M'='EuΎ xEfQ8>UPBŋS F7!v) M?<+6S"UĺIa;v98ͅ˅U﫡ˮ-֝mCZ5<iT(jo^uV h6>]w&8`~Ob\ō"JF 8#%hm`RCU'*6\\#k;Y9?´q)e x32lab*4~5ʡɳ5r7l9u 3fǻY&c>Jjٹf'Sy<=ddM!JCШ<)>{q:sq3Z'>SqR:yDl~s yq3hӔx"o+QK*; Wz-,*KTӨ,d / ht6,$Q93YN*jֻ D(\邰.EBf`*CV2pg`pZi*rF,n[!QY3vܘ3|޷W3 Lo;'Rc Mk 8- Jg1>$ Px%(ӊ-(/61ؠΉ SĀ *Tc( HFx XϜGǬAPFxl^iM}CUP ZhayAe;g6|6 o/hW~9 b b /0lq]z)a9 `N56C^ |,H!j[D|@0D4YV̢1σF5lϤcvoG~&đGhv}5/f^{=Lw̷tw="~P{& YnKsu{UKė1hniju[JrapIym Wtz92Y!Y-fޗ=t "H~XPE-ѕN|0^>s9"- V=rc'1eVlLE =Ӂ%)^;uҚ.B@' }}-ʇnR'6P )o {@/O_5CMd=WGm T2jA@͆I8P* E]jd2FiufOYꂽ~>t%=u¦۪AQnE %uuK?S=^5DІ v9Z.H+OV\#lm߇kZU ' {ŚpR 0G-h:]% ?CADo;8MI9U %|<'KNThuFŨ΁f1jK& 4kh, S2÷S1 $]U?v_~< O]ÿzy"(Ǝ4 >>AT>zz_;N.Y,M: @i$d#KaLQn+2U|К(@[ (qƣ IĥɖM*J5q,y A GSwh*/rIdeA2 j.(>k+/ CKk$U!Q' Tҹ.[&%S˖貌L;o^֫na$4M4parMG@xzAc,MׅeY+ HǦ)¡"?us>>g}hZ j9 #P{4B7\a7Ph'@&C$>G F&-oAdǙ}vgi}s.Z0mkm#G/v*>CΙ`Xl2XA@Yr$9sm=lH@bE٬⯊&&J́5F I|BBzP ʬhKρ$(bS"VM&uYrI91e6,#g^\P.Ej7:Z"؃M x$+R^ qE .dH P$,jt$>1CdHn27ael>jTc--󘳜1m)B .ke4JH!*餹"$p4w<崖ZN{JN;j{ؐmvAQu89tYڳr ַrONt'p|g ?VP1R U5\v@qLE$3*VxHN(cP [ ~@aJCT҂Wũ2681U(Xx(p~J18bdc$:gωj.t"|;j!v_v -Nj䀮dw:ޅYI8Da"X.R&]\Nռ,Z,^42/ğ_C|l 2ڋ U^[>+ʰ]4Tg @h+tJyXFkcN0ĮR]";KE :CL%q2"lyQRZ,`\F4s͆qQ:"@tYqUJ Z]`OALgiȕN;*[i4`8 ܟMx r/|LvUVLNbTn6m(—KqeYG2,A֠XYvIebMPd }Qr{eet DRYT6.}ͥaqWA.~"a.:z񗧫뛧 74:{f{f =Yѓ c7nzo<xsee\w;Ǽ͚*#{Np߫hOozM Ҝ趵nvZs.sǯun/9jͭ2X, ژnDvo\Ho04g"TtU_cIKɶiuVw?T]GƲaфKe= qβ h-=zB JZT(*emgJ̅+42}Thݽ1rcu9Ni~~98Ew2XKdl\[xX7V`=!%I(Wa8I*TI`tr;9=mr$7w9$B-=1UFwWh{G_Ns[Vܓf N+wD~Ѕ`ǏImG9j;*kT~@'MJhIe  H ~tQm+U%Y,IsV(^N@(DG꼹vI SG<-O03FA>с4"h :mFD%+)r`L( :H( 0IG M:TTlcՓqxX*0`6LԦLk~E.$S:7bC:ᖔbKg3.8d'K/Gk'Ƨw*^\늱b"3발˦)?,M_ I~N9{Ox?>%ǨQF5?ƍƷ;x~yF/o{׿T}W?yυ}ߧv ,+IЏ&_"GpOGײiҪm&G=mMyEs5IW|}|}ڿ,5rF>5kS9؉ӯD_8i7";^ZX?eШH)PE!ZIsF,|Y(@RBxo[2!vE;I,J喴}YR@R# ż t=b1U{0.|W5dl`[o؎A@ cw sv Ơu܏˥#n:ٍ"P\~{1q[[}Nw]şS}fFmr c*(Ǫ ?{;[ܛ ҝ~h@ulW)K$Y%hzq04F&*Nm7"__|b@ʂwM@u$gX[t1OE/*䌗i $/Q@ʣUPB֗1Sln-iqmu~Zz>|-nOVcIF>jyS)Ti[)UW3aM2 |We@ˀ 8lPk9ɥ.,B`٧lxb,Bm''#b0Qb"IEsNYḋȒ$!g/|fAV&EŕYk!8'{W+}$j9§pe9ӭ/v6"GA# њɡ11:6 8FW"G1ôQ9R4V2o#;Qw Гw/G6[a>K(v{5m 17PVӄL:IW\o( \P +oXje WxU[z!-XJ Qd!T EvQ=bUqE99n6U.XbL_H}?{WHr /}(uޑ9fg  yJtS[!RLV5:X̨/"P  {RBZI=.#Z(If&BJ-NsFjij{^g0$B&TRF#!&%UD T%L{ A:Ni(f .͔<),xI HJc'2BG~'G뼂P7#c$Gb%hZ.YC`)lLC6M2Pa;oI|եǐo<@gg\\%Q@8qeP#"Gx;k,KJo{AG;:w܁7VVm/Z ejk{L%Ȯ;,+gZMOnӫ ?G_ 鱙kF/wa)9_l#^XyTo3kIks>CMv4Wk׋ty>XnV˶ԻSBkg5\5jZNT ߡ^Q=z Y:8ݧ>BgĬuYc5sxF/jnoZً(;4\pI`IMm5muMY(X;{}}=u~yqbl]Wo}ӏgPi*ìȲo:i%ߜ!;d9j &~ݿpmf-c0" j>ԫ-g™cvy9ܰ;mf)ޘ) cbjjj']DRKM ޘ" i2.Cj2㴓rmr.NҽƷ@-w/j40:I{G\əiʏea& L\}qohNmɎ19ߓ.6øۜ^יLR.tEڽ;{~;1jLfDO!zا--|zNY3%IE_ngD+BAQ|[qR7}hS67T

)PB/aPRzIsUbUPZ3+a.h~bV/{c}Yޥ /e[ѵ:[Ui.//o +.;%$Rq]%Q!U٦)x&t8ꛪuu >W:oUmׁ  WVm.7WT& jdr56H&/G+k͔Q{zuC^/}lm;f3l JZm$TBYiQ#H(|m(t𳱆hy!^,"%᩷~V68vko՝;oW$kr9Kb1*&ijb NP*D, $S69kEk_cXa_o}y@2)JEN6#@%i0{AULGIY~ a(G2E2>ʤ9,i)Dן>Ֆ[g_6.npmE$A6 |ȓU!0Z\ 3rCT~&#]H3Et #{4|4\ܫ,EQDMƢq5FEauhT72Ε;v4R(It4*J2XグB"D+H4+ mr #S <`5d[qgx=BgXCx us\DS4Ɂ9@cN{oQ>&qT3Z9@\j&x\.N/hs. Ih4Rs9{/yk|KzTt Ȕ}JΦ;{}-'j~# emDw9^Iy%QXbip_\PQr(eI FHuINWgW3PH\C}HJ9"ZdBnqc" ƣ]*/ͬ{պJ HQQzև@Mc"`W@!\[Mw"4jz+մ#Qa C W ++rI\+:as&X6odHVJE 0ROY[BTQA*#߇GT+Qke1Qy{#?sBnddG{?3W  Zs${Dn7Rb㈒iw7efuz7_ꃅSւdJ\6D}|7 !Bt0{݅S.YKh#ꊻ,ƟTs[{ݽTu]nUp)gKb0 wymWEՆ{֍?rqqGAHL+9t4 kƘiV7OȈ!G̫5h8xZ,y},Y>!fmxV@&gظ<'(R/}77оc8(eP{?vO_~?2ӏߣ4܎u2 ?JN oLKM-Z3,g^]}>F=˾ڠn_z{0?ӢBv}Xȟ>k; $~v6ԟ >kABX}EF qh ϯiQoau]ik3 CiϤ JpP)y֨|s&DḅT:\̫j>Qrr}wz5 '"BɄZt;Ev&qКQ=8Oq㝖G[.O-7:G[ˋ\2ι gR>i_]*Mqf+E )Վ;7}z7mՕ#IBONwW}3ƎSM$7XHxLTL*YEVAiY*e2&xN}ffXc0tHVDt- .1It41\-Qkxqg3 D ]vZ 8a ?~&l;]>yԧ2X|3sL?F'l:FkRJe,DkRM, Fom9.ǫ}fjM![:g\4+(vVDv͉́կ Eܷ,zp%ys<Ԉp2 +R ;v$DA Mu63BG_+}fM)!HNphQo(c41hڃu 9p/R L8Cʂ"j+̸UOΈL*P8;~{_^76ooϭ>vZe2 Oӛ٘%toGP Kh@*-ąƱSSj&]Y$0ESd;eLY9, lyD*Z|5q[8qa0.־:My36qEY*L8CVReRJ\;YIRO/]師"f⮊칕Vt+0`%+r}1+饸"Fv+c.kR_ouKUeVb}6i G;)Qozt64lbH8Ga~GTksK%.Ӵ?;7ޛahPAf;aZLz8LaH%UA?( (@A~)RKܮ5)mp̛C~읫Zm z\xe,xO2OX-crpհD0z* \-h,N4j*m3%BOYE 8c 5y2c7 Cd~ \ ET+s##A&%^zNWr{k4PsA\Xȋ䊸i9HHlɽHNA TĕӃH"BqWR34(Jف޿ߗa~*c,%%bFapAI-%ǽPʴ1Qh%Ƽ±qŢg 2Q0&Ӥ,C;h C3j7Y<}4]L>Ϻ ,]4og1򩙛?Ǽm\2Ev}| o~pSt1-j Q.%J&) G]fɽ*>mՖQT}:Ql|-IA,4TEeIx!&霙,K:Hk9Mi~* 1ΙE)MZZi 6ŧ`T>ޗ4Mq\ٳ!^o8ɍ>Lg]gp M149;ןW{|^<'<5Yڳ^_g~W 1yrJF}^Bg:V.y4ZZJFd) ׄ7(+* Jg:PkAU^i,?"Xv+tOV+ta@5f#4q?3 &叹;u|.h6iЕփe_:WAof^qv:n~7­Ғȸv~?:8~J[%R fɗuvYѨϞ`7e 䡁opR<N,Sy 3A-9{Dr4zqp嚫cs)ZjY}yg5Ԏڮދ -1q7mXR&351ey 7/ KMpm A?mq3?[.>iė0h\ב==ߓS[E|˩ťwihw6\5XtBP҅T&`)*qي cUmК(@[ (qƣ Iĥ KTr4Be[&ΎBRRwh*/1Mct0VK.'\P52QJ}@{f)6K܅CGs5x\ T<.{&!S˖貌L;oV:/)F1Hall &hP.L.>>hTP/hBF',p:X8~Oh[{mAͿuhn O(bh$kNi< Di( !T#^#wT-qy{Mדs.Z0mRE:ihm̛D::l` 84`Psa5q&Wx.~G#vqƴGQ4`l$fLD <?JiLIsit #>^X:iɾ~QU_$+d)AJ͸Jg1U ʥ6;n#3T0TL[HӭʥHFg}F9t>VrY\;8;-9,3:x;g5bf*XZP'7~|| ~=)?[̥j5* O^j1'NL༤1PTQK*; Wzt= .6Ll0d )ht6,$Q93YN*jnJɸa]4K Td40ϬAؔϵT =B$XHzߏ@D/9A_tu\a$#V:+AVhA xϬyli]ұ3ڥcgKx FZf CB$#H7k#cVf (#< 6 vCUP zh_)zAe'|{nt[lPj} >:y"I#^c'-Ƭ89wmm~]lF[@N6  aU"D IV翟")/&[L*([6KvB 4ɠ$ 4y#JY"h" $'~QxG>/|S[6-@guﺹ|D%mwʎŭ@h}׹k=@2O{uϭItG:Yonlb[Nl}C͝W[W_[(RJ˭͏dy+zY^I?kvOH,?o 8 J nmv?U]B"A蹏3I" $i\Y۝ xRpENDLjM<(2#,Z#ՁHOi9Zٶ{clipl ~{;~xn.G;EfhPػդ?$eci^@#"9mip1 22ar>n\B{w|eIN1á}.g f'/Tv{ŠoT!GD%J'AܮX.|:$Mqn8NY̺UdZ$5E):| j}HQ 4&I "V?}a$e2RaRYϥ\K%9,G72RÌ`S`.Y M*lDSRR3TiuD!љ \/ & p/vn~Dž\p`dshu{aja}Dk.G2;/e0$5U>0gunatk}ۻu3y2S_(yUq">vdvj4s0J('\\*9ŽF/Fc7g7hCFHt6#(OԌpJqS^E?S*?f2{N?~)=uo}֖%oJICơRA" ujCN_hױԳG?PW\e15ʕ#3U;͈Kwݿ&w~*o|w7.gњr6#~S{y5[nvQKݸgۋo!u#=HVtS7 ƘadRF Y̚|O'z0f;^pQY?!FyVLu7:G矃]ߍP~N3}ÇOOӇCq9v- $YP  _w7547kSW͸G1 Jgd ?v~]?|\a~ZɛNZs[, 0 l~3u|EI?Bk4 ш!?ݸk.[kmT'" %=Rp%S: 4 $r"1 pd'lk(0yE>h?qc(Bb\}ZRz@{*zP2N#;8= 䍭'U _ܬ~{sQM]ڨL+pUt&ɥ&{ RޮZ|GgcbBH]hM Pb3҃e(ju)D2rskq $s *N$MC*&h fG"&rhΒ{Mk0 epBR?8.3Qn5>/uL8/^L͡cGO8ƅ@\*B4^IR*A2~xC6a{H\iNB˼0oNw=^Ӕ| V*]*߶͑#bgTV JMYI$r.e%BTJٖ|e% EMH\!W\fE\ej9uql q|B>ډ`J-v#Y\F-WQ9'TOB\ĕjS_}6-O܉`وL$UN^\e* ͷՈ+N@hqF AF\er[y=_\e*l+^Dg$2ML"ZJɩL%z (!|W*+wp*SWQ\={sb$M2hcvdrD;OƓg/Tu(~p!P'?L?#1gg#3lPetRVLB1(#s_w/E3ϋɅ_~ۻNϳgi~7HA! Yo:s߾Zp+Ub*8qVə@A q-̧R:Zq %ϙI}2&aŸUaWbfc=eAn o/7۱Ϟ%9Y(}42Eej|՗4*UgbnC e8|LQB!aC1_8^d{tUc'WV>#}fx^\^f}TdG󗐡U]pQoʠav8}5^-|_!/?\z(fҥju|#?ގ֧պE;;^t?埛?n ={RfN|Iގ7Y sx7!WXVVXMu2"[T~˗3[lxdYZWoX-S)֖ {2PbOIvQ:lSa$9ej379<@К?*I'UIZp|~YOJ l:yYuĪMs&j/Y 96jǬ;HP%ZQuz19>{0N_X-h! 4^^2_h≆*Rhhrӱh?'qwV,i:A2LI' ":X<ՑJo!c16*:%cCQO '+1UP0:ֈT9脗hU6M@I\[;%C(碃qa\^(% II 0z?U/]p%WYΠx$w [||WI_\S@,D!Raօ - 4D)kSqU܏[=ZkaE U9 :9Z)$P=ɩ`5#I9Q@Z5!&ji.*/ >i 8&NŹ%/`5}y?47<*SdI; _n*@ϖ|Zآ##36PM8G%oL.U) `p-/3ME'pL0s*%=%}BmfidR ))[}sbe H?BXC#w&S Z\OaPߺ81ժ پ{*%,F wd-:_b3vYg~Ab?v,~'}eHݫ~M@Z[3o.yXCCztAq/2b~v%=Q, yoޑfAе}bnf0<RH!xxkf4B I(Qxh"&k)|]rnom rF,|S TGTVsH|y`X C $Rp]V, gl Z(k{-zԋ*T\{NRU6F i&.x뜡BD Dy &g TrL6Vueu_ԸYvZiR0"'3(R`Ih ^kLGـUAaHG%2E2>B 2O}F1B *Ѿզ̢rX)"ÜTҖnO%QcsrW4$'[fL.7f-{*$K&Rٻ6$W m̔!Xxdژn|iG6E$%[*^:x*rS-V13/"3#EW*9;'O6z L{-2µ N(CV# FvC\#b \elJDl =blb‚*H$h0b7&bN2.w{3'X]&ؘ欐F(i4`Qpf29'p1g5k' h8R22C*Њ4bMfB8$Q ڹ1qaeԏbig`Dll~aD{D#!jWqSrCn@2OD 9@em6h5 6)MyԂ3>`(PTs0!%̈́NSeΆ1q#3Q]u66KVE0.=.qqތwF!skNXH QpHQ ;.h/\&pih,G3}Ӊ4xVfq.]A$IJ#h$J")"@=Ƙ14dF]dMaƝۼu{[ ^w">ۢ>9 <ן4~<pٮ>Օ_y~?y/vE ?z Ĭ]w]G(/{8q`kvtqRʩ;0@$SdC@Zn3]CuH?~׀KWY> zw~:ԝ<B=L h7iqCSԅR٘ F5zZ&(~gpk= wo:%7H{{;,ptXl,r:CϘqK-vGo\\4`8ZFsR /]P)\`FtK-5˷ԶS6SqFKaTBˤ➃a՝ 6V&6[Bk6-Қڈъ]@$z6"^yݴy?Ξe~xu".[[D96M8Cj\)t ev4KC4h$FJ= :T$aED4N'Qjfrm,΃!͟{ٸ͞T$x.(z /_tsGt&B J"Y^LKl( )sٛVvoWo:~gnYo8;%O^^\DSoΟnwNy3Lڼ̜_ )k=LzN*5)_/vm&s}{嶷ͷd nk6*O-z'(\ubFzI©Rɘw~-=Ro喨FtF:9#]쐑.vHAk-)X.sTJԸc&8UY@89Z G1QqTDVEFFg}NH|d(,MD w@_zbDs~piԚ:q{c \ո-M"J6̶g-<=ו[MYΒBKku&g=^m~s\rwOG3FqK"$Z zHR>0RV 8`]I{JljX2&݉Cj)( rI+)߇s߾>mρ8RW5B;șHD_\CZrw>YARq;좲Zy>uq* _*$:>xLJ/!Ch*G4Xqz׊ S fm]ԯg7#Y~ ?8&gw1r6}kuR[n'};(©y{9Rθsy.UTo$5ϪͮڙLO: s/敼i_:?ui#bf̯uYhF%ғt$W,q >W*GҢ[oy1,T:PsC]XS?.B O髓__oO_zJ9=/p {`!})z>JJ!MUF,Sl_^dg{W}`_eжxmo/:erQj'o:MVz v"bW[" -c!9jD`_Uލ[>mztM<,#+?L`Ǡ4JpR Ϊ|9"1 ^*>$Pڸaye>wǶ!XdAk@X*)sA){iT2XN#;}\^LȦJon?>lK/ëgsI囩k {cP}8I 4DR 9(=gT0V1=*L.T\0tcarHAjYNM$%I@Ht#:'5J6IQhH$\Yjj\"jO%ig1MBLu@C;W(ؒWjm 7y|Ӈw'[#h`1ZrUiLh,Xr) ,?ĵ$]]kؚ5#+ [{0T,1-7*|Hspj$ȭχQ&,J.Emn6k(ArMt2x`x`LEB&Q{<18s$PJ|A04$jf1g"ng0Q*tY@sH4ٓկ}4yѨ)T튽eRKBVZTr+]ϭWϭs+%asezl&ہI9p'<VzʸU ePd*S,2W>LJ?#\1& 5O<}2*KkUҪ=\Js6sP|7۩nCJBЯAt{:+Hjn\ X}xg)cCv֗Kx츨u跪j70/\tʱWlr5I)2okYZaIiuPJ7/~YM|ug8c^@'x*|,gQ5_eB }Żn֎^ EH4'+d{q;^!]Oᷓ_8?$B(R8ԩ_xUߍz =5k:p.5GuǽqOLsTfD׺B@GŢFO.LgAnԅ\x]helLBsdXWاBŲJ:RjbB*Ƶ̧b{ED uaz٨% ޗ+'EW$X&,g1"[M:\0LJ(!DwNSlk:+5alO\ZT;.|i{ˏJhH9GB~SoeB3PH$hYe*Zqk|(uR[u,c1Sι8 @E*E [ `oG nJ#)iO45>6w|U-_JuTj_}j[P< (lr'pOiPYbyЂEg+fh8AfX[~b?M6Y0ռ9{0) Fx0@3){0`= >\X4,2Z,S{-%SܫZ<%@nwPWCaoцJ䪔h+̍^MfJ֔&Ѹj6;mU9y:Ţc*!IȒoӜ8p(9L.S R}JN00@PiԽ5w)z>}R$ׄIq۹]Yق炝9 3Ư..;9g\̦hE羽s@uBj:g5$ɅrrjNl٪RM {lo$3_g~o{@X4k8cR}ߋGvl ?݃qI=퇏F8IrY7;/;ϗIRʩ`LҫL;•Bv'cЮrU~py+i 9V7ɐN) ,8d.u2 2ET_Jj-ǢQC?Tk__w|œM_z8ꬃI?/Pp1u^TyCT._`tiͽI\<5_0lGheBЈJbŜ"l 8 )e^F|˭o +Pp09O/:uMpb&t5'I8M1MjalB1CIGz{ڂ Y4 LWYh|(6 tʉȉ8:ȥQ3A$mTҳHbPijխrHcz߽>d+={s{9=z,\tt>@aiyl;(cP~jn](>u,9Y~d@*)PL.Ԋ=T6)@=*7Ek}fJF}tj(hgF\9lńNa,Xuib Sӥ^I׆~Ѿw%ѕ=p-k`POyN@P1HT4\9G`Ta6HdR'I??ݪE+ovR^YV p.3(R`k0JZPt 1@P$Ho!ZGPne 4DpJ .FfB5@ "* W*޷{`!< ,8**7F`} u=}o4WT vwQ1o569yo/OT+ eg"UN UQ%|;0BbTQjSH#EZƠSV A2 MJkbl֌Qʳb.du!kuAuጢLڐ`hoB6\*4`=^3fr)㴬mT֣-dDъ,*5!/iKBx4>,Ff}X/q0 1+cшcW kDjV#4!^ŭ>$b"61G$JE5@Tm(8)<(AGKrŬs95EĥvDYKvՋ^^l:!)FpA9"8\8ՌB$YmTzzXa18}(wӇPa G#7zO>Xt)\o:#PkCQ)UJ6eq(/2A%*퉋A ˽՚UW9hE'6r>s;54r:M#-mݦO7I0ajv4m@dLBT%鹂O _ I+~]h'%+fcŝK+.DO:G~^ph)9h'IqxD$A!#CzAJע"(5t Ž #_a3`V9S-tn\_L8hQ;ܐb>U8ec:?JxҼ0P[qL Dh`"{+Zx*>>xI^ң M stG Zy.#Ji 4L`.IH Y^3 ߁|Mn36.}μ\Kڰ?Qo:Yϐ Sx^`|·緫 74:{gҽtgn߬i|ͮ-߿Iڜ-Z{s=?M61 o[]ZI6^{8A恏F M9Ыww>ן/>mg -6.p"5\EU$ZFz>IJO~(TM3!2H?ΏHgGg#2Q+EA1:8 92% 4'X4Nȭ -תM9̈́"£&Ȕ2` r--*#(,Y*^FD/}͍gl+( OゞOkaj Es41-^} Q|yeD_o5.0wEt.URq*IepUDU/ \B|t5ۑB:훹tA` Guyl5u;m"*蜄`S٠D@h"9'!K!e< Cb1j(C(]PR0% bly$63S\Lɥpkorgvn]qQG]n?jX n&JCNһ(Pތ$jIU!&& JP>FVD!N|S#B|wo9~IN L7U2RKū4QhojP"EA,1GAcfc/5_nUǻεdpf_O7iDGHEF-rIcgC?opjjJ՜Շ˜&owƯg7#vs䃷!1Z3b^^~-3s;(BVv3뽽vc5'a/Ɪ|nf+uGA?>xǰQE`My_.{y}}۟y}/߼{~|u30>װl$AM?"hAr uRp ɚmp+U&͆8+Lt{ո;|`2 E|>r-֓D%NZ 'P,F9u,Hə<L<&څh 0Nej$FYD[p0~,FΖg?g}F׻D; ]TT|Djm#GEto00/a{HINw1%[r,ɱ([vZlTԺ4P mVS\c62I{\_uuF@Lr<}#2& 乵HO*Vp5sv{y+E,o r9bWSڦFuO1Sm+gLe0o7*u<(,g)yaQfϥ\I<=ӫ˃š D5S C|m 9)`;:r OiU˂?9F&XfV~!!ט1MFD&Jcǔ;68ͱiǚ3̵ua[7 ju`o"L4` BK'/~L]qN>Ol 0#j3_kR{/+ [)4/z dY@`ϳ~[3'OON 8( /4L}3ttmn;d |-?- Rcnp^!`4\Nj jyvtٟ̯ F?u4ĵ\ ZL7w_2,|38\MoL-|EmA v 9stEZrR>^@#_AiםOFGNy%N uuYhA SSbq`=A=ߐQ-WaZã(Kd Rud5I*&92r-+sQ~ >0vy!7" Ť8zƣ9++zL0˛h=ggOb_oGIɇ|5wmL'-%Fsܼ&}p]?Zr[;_%6vKfHl CeGU+|L35\'_+%oRg |JhsT2NtHk@W6PA+"~tPqta|Gc9nw?_`wqtK1,*.*Ʌ\J*)~"%_i'G>n'g9?.qlɌ|n;Ct'.7Y6%ަOYzۨ-u`Q[L4DŽuB%h9)h B> Pb7`h 5֢"rF+ Bi&tN1bAVW}{A}f[)ˌH}8{X=P܆_<,T6XDg<܋ERXq4$@xMQ"rqYjV 75J3bc!kC#Bm@`>)/--tYKhI+bDx)nSy _ᢌ; ό/~ V{%bKdLϟCf34t;'"a-4oԛGxDbÜ}(A ewڞ96#+ ^&ѷl>kw2E^4z6Zi0yDANn$>Ƚ ƫMңz}e2-=mq`l>MZwʠַ`<; r5hE[Og~uj6zrl6.-ؤ0>|0AVMBÎQɻb  f% [hI8__N]2BG5W㗏嗫%a||%*Epң&(sgb*T.ig68u',2GOgǶ_f{v(~9Br(/j %NZ9s5`2H \9iAlC*WF`ȧQ66U8|=|~7{IEbxX7J׷Y?-뙲,b XC+Њ5b XCnZ6SZ&$ߊ5b XC+Њ5b X:Y XClZVkhZƎ~XRS?oM7~_!#N5~ϏWi?}4tqS:oJ 04t 7&tބΛyC&tބΛy:oBM 7&tބΛy:oBM 7&tބΛy{ׄ_9M7stΛy9o:-\q阺GF6ק^~#"Xo7.GǙFИ$S:SbLD xtL IrZ*` 74 Ig寘<$4ID @q: B=Q'Z%.%Ģ8y!&Pi^}|hL0l<oJ}l97}{UE凨QȌY"$,GZƘQ %(MYCv0uՃ L_ȊYrAГ6f"!gBn\+ j#c5sPWFcPVBٰi`7 <Œ7 'Tg׳vs :q1zq>r`< KDʼ FIS WFX!hHu"$Ȭ7#!()M$ G.v \%L)`2bW3g7bdW\cAjvc/38}\wVi qL (!Б_W3}L3X>(i\ß?qXkYZ>4i^&ɲ-#Lg=X3&ȾwC|O @:'eЁBC~I|thx0 |-?-smT=6,ލ> OIhř=.`gGgWlw?:DƮ(써m($,oߌh4'WS//тj}=({8:Bw:=Ř֬ 4XUԎV{gءGѱL?ʵH}Vr}иr6YpdfAxd6`Ox5ѤF #l yܙU݊՚ )YmeMVِ_ K !Y`d-U͜]:Gn7;z>4_j>tػm/S{&Ò\q|jQ8:? THƭJ)ԾU &FFuƏ<)]R?4~Iij&peuJJd$Hd;HP&z\潱acUX !tD7 +Ra֕ OͷIe{4EoGD1Ah, S{7i!pS6y)-uy[yp|RGA/#&J 7FAi,{@¯8&fpf#J^Kl5^1x9q,5 ^,;.X~z.g_hG\Zq•gP{+ԴM4#Z5c}ݲsv[)gsaaOϦcblE8Ն\Mjwgv8A7{nX{7ƴv"86 G0b^GlNj>Ljh핑{=d[-ϪPYyG#t$w 8}1Wc_3F-[v oӛ=߿8~p???=Oxɻ;[\u#0> w_; @wjчh ϻqތV%ӽ-*%=Rp%S:  tILĘ09:\%PzU9}Qr}YRzrz* ]ɄZN6vqru7]͉`W7w >Ξ_*_az3L0N mr|3M1BUH] t*Ǚ$%Lƴ$}*IԸp'm^jNεw@IHH1tBUo;u1AHa.438+8.ՖDU%di׈Cs6r ?1?W3 G<77 >WKo5>Cc/x:Rƅ@\*B4^IR*A2A1o"Ҧ zWPm}]}ƫy:V1O$N2@WN- ڋDLjM<(2#,Z#ՁHOEt?v&Ά{? ?H\/.!n?v6ltBP٫ٴ͒wN/OjQW1as紥(J>Le}44VrWOUX1WY\#\k*Kɋz#pH* ,C1W(-tUs< siMA,bR.-\s%@;UXwBi%}7WYJz bzaOrq{^T:׽wK[ ¼\|Dk 7?{]r46 8|9Њׂ -9Lŷn0,0cLgio1Rb_+8 sRs(*K,Rbc.^`T ]qxU>fQџ႖ɍ>OfE1 $0*U^P iX!cNk N+oHu{*Ћ zT͝W?}7? G_-[LYu=ZgxdŇl{wxar]KVt籐!Qv[׫@",Szin:Ua[6Xያ}6ԝ Ak"ub<# #itx8ʌ#}aGalѴ=R*fd I8~2(Ra2B5z: o^>Zu WCW).T's]c=5hoYtwE JCA bg5cm2 zIqt|9;uQ7i#ArImy)\hj2ah!(Aօ_iO7ȅv҅Q/dI?fGYL>}Bv6Cu}0?lj7[o7m[xBzD\һOOyt7Ywx|=^Mu5@^w }GÖϫ!Q-l8瞫mIgN,PE~T5s{nmf1tq}gڦ;,Da~IclX.V ߽ll[IH/,1&0%CbcD0)hQ9boIGtt> 0< }+1i3d|*zJ 7IS.J Ar* D:Vub +VXCtL~۪!(relʆ)at|yU571S h ٧½ _λdky>행4Z=w@ <? USgv9$26R"@̔mcmc X%Z)3)4 I\DĤeJb#ol'΍{T7`r5I^#ꔒJAT] -,0vGEgO> %P ]H=3:}WOHh:VAI;|QON˰?;rc4bSyܸhϏEUTFY;l !:i`NHOtt|Y"WQS'D+,RkQUP(#'8>jK ^ P(5sf"Tؚ9[k]Oo3^R=,Co7 W*AnpPL䈭wX#0 :)3B%&'J(]&85LIBqTȠgқ- IIE ޯ$|]ґ0хT-#vklFl;+ZQgGQiiI,H{0Hp`NQnPf!FafY7E.q8+ǒbfE;$` r#`|`#Nu5ٌS߄'` "?vED2"{D\{Wa *L y3 0I b8&(ntV%yg($^2XobA% 9 Hٌw.V:[%hy=.F1p$8}FZI5L# RTލJEK H[C Ҳ_ _6(HVq5(X!:FQ}\@9XlTwU^sq.`9O'76BD;kHE>ג)E `! A+%͡y|xl-2S{βT})|IO&Y_Vaپ)v;XUq[ʧ@2dnt8π)K%:ͲoC%8}8jm~]wD.,]@CwFU8X7lXڬC-p6D,LK$kҞRR *5=qyI{ &uD90D3Eyg5;geϣ<Ɏ4ҪճemdY]O.p892ҵ8LPLKhgLsN'% ƓJVmpR}~R$'zLmHw -\',ًO)Dv3WWߟfk%em~V9Ei@,+^㬜#̙W'jr>Hw x<{jI9ռYkfqx6}a(FyβĖU57: jf&|gz; ?~U72NL*Q2HPl4ޘ݃;=3f2pykO_k/I‹A1ܿ/gȷ%HdF{3ApRY0Hx׻/-:-+̿~Ia.eJUԉ 16;"0Ρ 0 fz̝NZ&ƽ38X0>YaB. xwz}Ӹy Y>xR4/$iu7ir15I rIL^ȽVB 򺡃i+EnP/a_@]9-r͙A6ƒE ,=ugX(ymPjbH:1F}<(h$SQn_2ݏj}.]f0JʿkvQVPȕ(gF,`Kx:/J;J;J;J;Jǰ)0X {XhEN{a )Y`+,:_B?f|bek<펽6r.%WƂn=_W[&=< ԘpI_XwL&6VҜXK CLdԹ*8ioT`{t^NK \B'qaQ4)@i#6 B ̷-0'X_l_2%+PHp=nȺϺ=kM+{ye~cZuJ`bxW!uyz}PbC4뽕2CHCFYilU7g^]"煔d<^_tIy[I. : {ǣ:[ JYl$ڧmnqyM\]]^}󃈥[7puD%c ,Ibd墒OXr܁%mYbġ=\A WϴV3ԬSppUjǒjd\%s:JRuJV2+"*9JuJVR+*b*]cd-E]d%C=\Ab N \%s_ W`-ŝgWJү]EDj43R.>euW}Yaa'J8(;,C3DG'xZ%PCh;<84 #ZI޽ܜk6G`X=+޶3Ƿ# .?30wxZ5br J傾\)S}su %W2:Wb lIz|0 J`tb(շ~Ҝie~O7P4Oq-^97[ݬW#e뾑hf<y,![n54:O+گ%[eUMnnj(EǴ ~B*TENcYW?@&KA$(nPK*,  esF$&: 3feޕXMvv]K}<xTsab4˸ R[1ǘaTduyzy&d8r<Wی?2.]6{i^ڸ6{i^ڸ;=/9Gp$D/mkƽq/mKƽq/mS(B^ڸq/mKƽq/mKƽq/mKƽ|slR;/ނ6y"3"ܧa [KeT#]錄:9CN:넓NHjAy낦h0X SO0$ 8E`Fy z@::=V)n[$QA-m=3H 9-;魙IqK:A^4oX,5󜱵bj+1ybNV`9/Z3E-&c\XJpHXG"\JlB2 ǨU1kx | \KpGG%[^h͜-wg3k3|t H \SkyN0ݭ؛jo+쨥Gnt V4R&YL׍aiQ #}"+Lp}EBRXfx<6J/_AEyM|= HitL*"1Ʉ,XˆB |D=w<*-P$9.whCm?Mr?2 uHu" FG6N$# IjDIQ^kjf a%&34K#qTmJ('1,FґjwTr0L7s $(&&b1G"Jυ"ijh#Tp^~بBavF~>w[tJ9Ag?8Z$K[ iE"L]Vb@X/'( ƸYZ)՟K.͓r.yҧ̣}8iH Y+pZ:"a :RLݩ)pm@4 u^?k> 0! H ˥CRͥ`i ϵjԋhZOa J C_Ɣ@Ry[_Dggͼ "m ^TJT,$~ se9TlLuJvErՓy;?H3c͗(UR=ۅ"j& ]|4Y!M ~&Ɵ~אկzk;0}KLfhxTt1z0]S8R=Vֿ)䪷(A$e-lhHz1-b6SG8N V F n.;/??{ȍvmo8I6e/d >vdɱ4 _ղ%m6m˞&b7YůȪ~~}7o_~}?u#M "{{ =]Z]S%]5{ 5o߫ӕnV|3o<׳ veb\ncٽ4JSHlR*D+^ IjM[n\oQ#mkvzGmpDCVk0JZsTxgTJY( s~p}aeV}~5G"b;1JYL@޾,q>л4Zq&S5o!3OHP2ӻVN9.ޟs^L.6T lǥ+;Kxn5;q>;{mymyK5j|K`Fw^{poBLf$AF8h1 K;^4hrcG9FEC$@ 5X*1:qSz-L$1%e1S@8Ғj/|f)ZimjlIdf/m'z)xFvǤKq`qcŻu~.T X袵)9L;19VD,3>'GҠUCo+^ݴ P&[ַcX(vD֏oΔ>iE K;&N&aԪ)g;5.XlLQKla/ $Ʋe2")e= 8gBJ˧ImP2ץ/JQβ,9>dTd.^ Cb&2\eXM~΂ 9>nlO:!v{ 5m'dS/mcoL'.<DE }-琍DI~+#UU58:s)s{:]=V=1ғR|e &v g=. /g_yMyK6~b,ʳ$/=LEr]x%>RZƃ"2,%OM4,,0T|zF`{qyQD$xHlnRTt3)VdI]lI70|.KjOg`&t9)@8ہ*0t6{e*R\*%4y\V9+jV4 j 57LӴb~q ?bgeK0}gۨ9G(pVHQ#c,D~k oDd^[4eK4w.l1LX@g'0>8]!1^ Yc72k%# ,\Z4"1ɂd)6_SHXAHp=6<}ˁ;__ h9PkҟG\>hmG\zJyF)`VJ.8!i#{Hdji6^) 90 B2\`E.mݭX{ yd*$p!g*Abp.IbܒOʫ١%B)4k+=p˻8=8`gI~(܍mҭ『Iŝy0tD|68jucyY7k;sTc4#!Td :& -BT"X%%߁E{tkZށuFz`݃-ݏ|;<eRuacRq%t:gU chXS/D-=::}c]TyyZ+p4Y >tveHsB–`4<rkK#}~n6}7}P'GWG|xn5}jy'v~(?'mpקσ?#fJ^[Q[-{lo~zO qqӵ繉;NQl!858#G̀*wq~o?ϰ3{@u!d)cX=ehFBiLJ*ΔȩL?"!|'k'}Zw[]dnx<]݌FOߍL e?p.3~;_qVZZ=^~ZRܖ2:m@ 5%bў3*,A*dDh`" 0 Xl\E%%4r1Bev:#yrḶj^iHa[:쳖WaNH>VK.׳Ul2RDmu8dL@r9xQLgIrX.P*2Xl 5D:E˪PqEAӷON=?焆7F_R?)kGNU@UGBFQ'iF޼{ĘLL۟Wߵ}r7}u9.^㚚mRiTFsrJU1w 9>ʈ젤uJ Œfo .njdE';=j :+ ;f2)/d+Kyԛw˸+1c|qFח4`</wWEޫ?4鲒ƩwH{ZG t>*6n%P_ŴU>Xŏc(:ik5Of~x{Fci𯾬wC#g9ܔb Rj n #_qJѿ1o_^dq:>cyuY7Wџ0\?lOayh< :rN8OCj=!HW͛q?d Lj5pRjy^74ZwۋTԆ0ߔz_]fհmsp~@l>ŵ֝OloF~YRn^^r::r94ʚ(D@1.PGGSqg|^?˿?)W_~-~}+~Y}7׺}Zvbr/xgz'}H]t|6{}`?z u+fWfkny;] -,Y&+bԵ>ayiŨMMy+&0o/QVV' VrKz]e_[-BXp^n9SV4Z|lmݠq* A*ÝOH/p|JN()QeH)C3)79EpԋHg={~ي5eJ23u t&g'i=ҲNOƧ,'m2uN*,=xYøcN7%=!˚~<+QYwT~! Q8!_eH٨5xkZ(C+@YDht w)(.C>ǕB.2KeN Tޭ&Ύd7} `G:*{rr:dJcpY2"c"9eMVc2X@Ss&pC]DAg˧^DQR+$AP bRmٮLW9#Xg%طoO?KEqC,6AOe,9= i%FWj?櫡wBZ*B"B.(\˔26IaQ`1kN2ISՀi53)K.`ڦHT=&v{:b"VwETE,50-4+90O?{GJvW|'uzfϣ[wX(n_dU?nnS#cȬ_DƃR $2aAJ+!" )JDkJ.I 6+`DRIɄ@4;"2-KY/?N=$6:[%e(;Żv3!MJ^sNX Euvv!=7BNR0 \| -ݱ-P=L>[ѧRTk;q E\Bޏ2TSq깣G?TqL|`Xg4n} {oHȑ^Nލ>\eZt>^?&Mߋ0=\ ϸQkR; xvc7]_Ԕ{S Y]Y șڳo/nH zP`D'WDzCMWmfzf$b褪eXuM bzM燓ൠ 0PFtW4@~5IƮ̨Im}eH=ڍ4Knb8LryCzkHC' qETqUEPT'q~z+勋, +ˊQˈQv+N\=ӣ-ŪwH\-qEj]Wڧ6 <^\*Y'^bpvH\ΈB.qUbURN\Bq%,W`;`!W2X[ *U^V+$pbgU!W]WZ`N\Fq 3eeb>/}=}[yR:nqvhfx29fRxSC›˦i,eǿ#+ wPȲ}ІY&iڴ\ě=JHWH)kXvƒ3 [cIoHgNHG*pqEbm0PD|]57wt+{)9 uW+Nnt`pgN+Cj ¨7>)O|'ݿmJvɾXɆӯ{rިs}{7Y>&C\>e}t*Mjfg,dtv%o,l0xT){%43tVm7Ꮤ8.pR,Y[,2YH' yJmma76}`ǕSscwWT#Z㝋Wٷ* ]]xx{@%%7CE PxEZhU> LI&fyZpj 5 x) (Z0/!y2N<=2{Aa[{UIBW z` >zAeՀmT#sT(BWG1_bRh+\ KZXJ0W m1BdX+:xJKYmd[/VˈYEC*a4 zt4:+!JMٕj7`_l?">oz욮ǧ]mF䝤4] W]^OhȠ?! L+j]ϿwwOtnm߽uԲw{yx}c zky~w7y {~ 9|MѥAj|RƬbnkN?izw3zyԦZl9=>/:tsm5\%Dus\ua>c Jz©JXn~V8DC?.\I[-RV(I7Q)KN&Ӿ2r1$e2Yr$HBuzPQq!P'3ʹJLks :kTfcF#jpjVw_~7TAH-bV̺RbhԠrbBI :ilV ʁ2FPB_dhziFBe‚JK4s+'JkQ&(t/LhmʎZfd %#pɒah v㹜N昔aufA |ܮ 22)7/)9 > 4YlL-Lb|Ju0$1m`%AiTatBaz66LԶLk~E|y eؿF%GC@1NC@l?i9FtϮG$ә&'Vf:aΊ"G3(U?4nk3#X~i޼_<?x7nc r6%}9糓Or;[9CmWE?;v➥jiI#ڕoĈ1,XNAzУ٘''wL5Q'\5ꊹw6QH)K> l*W#_W,AsqKQ+:U'tvo^1m?G?o?2s>Kh.Sa% $(g${ ƛ ͍dh ɸ)6X5}B{gk@|m?~2 z|=\K٩܃r'BdOq5UAAX;;yE! 0 .\'[ڼMNiM,ѓfEo)J!߄L`Ǡ4Jpr& RU  x/UBϮS0yU>_ N.OyB1dJ#-\yb^sxzCƘJf֢ibg..}gƆĂ645,^ߺmA@ cw}iz}Ƥ\0T,G9X[I$ҩDPZ#Wg)!5&XIuHG@-rΟ $!FS. h֜j\ڣtDs];m;Pk䬇I$o$P0qHv(Ip~s;Wٖz^rc,uc6+AZ& QXBAf 1o+;FoĻpx6`oƾ J$bo^J1 )PBڙ";X4]MXdǀsiȍ$/_g6 2\$?+ٱceN:HAgeFv 5}YCtcH'Ǚ 9!$qv2<6 2Ee傩*K& `ieu5r6z*:j\#I8|q{d+jGDzZdb.6ss߂:{>DHiý7OI-Ċ~1Xi.xBr[v*[`Qfo*YDeCG9, l =frr5#y#\99x)d0̫|vyw%[S4V! ah8ʧIV2 xRiRAeYrUPyf)(RK䖧/2=GFo.mGAJ1ˌ< E@TR3U1v$d@Kh H$%,1%c@b`Dt)!x 9ďbpv4)kAX'IҸtⶄjFab,V.Qc+ץi1mqpbd.K2S}geA"hF j2͂L >:Nf Bb$Im 6:'-+xky "d5߲U5r6(3qC }#JՕ&u͞~[<^+\3$;{a< X&]{8vWRⓡ#:&F4G5o4Y2 BSƟG"d_7k7%nPk|:`4nj[(y`4Uk=wy6˵8]-EzeG;TV5V^{"U lbBzy$[<}mT*"%HJk@Hh#Xp^*vp~~b\D`jA["0\$&jRZ%&+"e (HǀFfϩ?YQ2֟'{dբtH۠~۠1!J(EAuRmCYq+~Z>|;ԝ|ii'A(Qui!J{JS">k=#< #dYnĸ'xRR)DžpZ9G&I+vZzHB $yr#\p=xφpm=kҋ)jlԟX݆Xb{5M j<Ď,ęF*B4D$|RXRy#WITX'SF1ZjLPib#XVr-BĉXo#9<=q-׷C>~C |AYG+4P'.AIWBRQ%8D-Nd]ƾ󿯯ߟGLwlz-eF#l }+衙wwO9S*(SUUuMЕM/6yTi5ӳF?t@*h/4i)^񼱧8>vHMr>\C>-3E!4f{D[^;Mst![Cug@MG\hcC5eb;*2aԘUi-&zZU6@e-M ]~羻>9j5x³K[An҃I$"8(BbaI4Ҩ1{&6j[ n/x?``4i P:ǤU%ar WDPb@ OI tXPHNpIsjՓ^+%ЧӲ`} m4&oM6ֶnmt^﵁ZE๪HJ"P' xO,e{࡜M·mDą@P\Xd%pJ^kfE4 SR׌Fi!]h 9ι&g bN۔-k!r2pRګTORI͉~ iƝe)9kdilHem ^zVq iyWخ^구]#n^麦t )]mi=47o^}qQPvXSd3݂%Ֆ RjH ˆR =,VreXb [ yRGuʉDyf?$5DVI%=`L-Q1HL451sD Dy &g IlmmkpvɁE^8f/í+&CpZ L ʆ#!,P25jJTa1Q3 E4e'bU T:RD@TK$;*k塸X&']w&k Ik_~<erVahcwzW5UӅhopy+1Ze]bO D(Ta*p݉S1 '$d1V Iq!qʐz^ q,tJ`QTrKf͌*qakq_.-s!я-Y3W_Bܝ%xz9M;.Pp9~挭yЄ\#i@2&Rfښtlhvf܂d1j힋'nƞOv,&ZuѶ#$&D-3vkpfl7PX5Y`;Ilڛ,bxJ XO\+u3f5-'HB*p3 6 hE{rMKG(ºD ꨬm[s7VFP&Ca㾌-3"3ABګ:: ޢn@YC"VԈ #Q*#MI>qIRQy5D%UEX4,[CCy?>| #OZ# kح(*zǧ~4~lB;*@*W?E?<"m|xxë|r$yv~+A S1Ǐﲗ|oؿah\C)c7p^Tdl\=p#WWR,A|<^^7NEu(j_#ҡDv m;n.[/xN^vvuWjICUuJ)O+*]<`1PE9|2{(}QOlnQ ,[+QX\d|<)U?^}6lǡgGrMt`^hy"%=0b{CO:DWX1C !BUHW*OWJ +N cCt9B1(%鵫HW1]+8 ]e BW G( KŁ\GHh"+o]w-pɃ0i(O0uFc"(~hHrBUUBbz9yM29>[B3`M|ݛyF~Q*DҜp%*֪3tpЮUF١UFkW/ gu^! Wv^R+# topz;]'>AZ&CǓAW=B.Uhg*դ+tqt( ^ ]1!^ U+iW*}/+5!HW Ct+p5t2Z"HW.iWhW1J~tQtrJ0CW|X '97 )'hh7oP΀;dgBu3ڧ7iEҴ4BTZ ZLuŹGr9g޼{s*Egf*KJ> +R:Xpھ?^/(O8~`g/Ex5yߍC5?Y_Vh_KaVEr.^=ak1ޟ0_{$٠lɰkO:{⺸>m-'V}}UY\_2ys HI^[Xb{5yoFW[5ОE =}ģg# m#T RXRy$g$*,TRb{ixƓYg#ܘX-A}&(4,GH fI[SsP!Dg*yZa[?rۈK]0^ +:Zδl&J7?ٻ6r$W|;L;,S@;wagƲ8 -Y˔-g-ͮf*V ? *:QnֻAfR5cvWQ8d2MGҚm R -t DžR#!Hx㏇z2N.#;G{Of4 g74'[64 ſV<_ R6Ch*{c'{&}2g+~}. '($xdbgۻ*URWBZaBgv[7I?H^\_ ά ~AAOONƓӷQϿ z]9Ք`=/L2u 7tJrCp J^ٱ%k k2d|uBLj8׏~EoGgu^߆{YRfPY]ŐJpQi{4>uM \G{\s4< >u`vs=cX}`W!pZr.*?_q iylpzVv0l%٥|a0XڮqbgmkzNGxwwӿ}is?!kZԋM3uU|ak6|>}q\3~^4@ WUZ}ږϜHnxBv@ 0널[!t\N6MF1!w$#:M#ȼmThɭxh<(.{T,nՔZY ״uF0GFcpqTZ*b04tXPH0IsjՁ^=ЧӪ6>֪5+fr|9|^lem~yc噝.ENqduX w7]q ~;ړ6qɽWBaaZ]Gw6G:pfZsݷἹyDwl6NQWvWw@p\NtD"/kD칌Ķ,#=!^V<5Oޮ9^5w-\_Y͡ޅ0$D3qׇJ&}.9*cGPȲ>D%Jz(, £mAza,JU=I& #(!Bz` )Ĩ5%Fʄ T8۩o٪po_+ߟ-Y$O%4OMқo'5׫weqgJSc!QE!J%"5711&@VN,+x&P H  gN2jeT0!+1>FI-$1aF:BʣkmQ7s6TQ/"!y/F MRB1ԹʼjA rz"7U"".Vdh 6:#rqzdnqQJp/"2rR\+% HL I֍#caXPBֱl` ́/da9LY9OZWQm$\. }eFO>ʧދ`h@JB∴ Z)=TJS*Y3'm,$ygLD;VirjV.Lui5 Y" U➘ gDuSXSX9~wiG'*^۱W\l'^.HN\(FGτV()58"s&qL'47`hRXK̽gC.dum~] y8Ͱ\3qNMIz56r}x]ȺW߼ף5jYoAďcO[X?yG̡Ss3$MLv>k竕e6E^5]e km0wTe$ v0tMßWZ%j6*]g$TF ^yPnE8 "w5:VϾ.W6zxtji]mJ;JB c(8CKD0Ib nYR#l r:]YV;Wo6+\X֮5ݮ#]wH}82ҵm,P730I*`Aq2θQF E=jҔ+fp3 #b1q#YWMY싋0.;\)Trg!\P ǴQP4Z8˸ǂzDZa[vkvȭW wh(j:~d8|,nCl8p$ >%,FVTݺ-qV$o'=qtO 0ϕ>w:l<" FTcį ‚QȤQIUb'Hb.1a=36p9aIdI"Qxp`>QP%a.eN1y`6>xiR}\u{\uOy!ÅPzA}??(ʑ3$7?ga0 Կ߽iE޽ NJM\}?Ӥ@&\ V3d0,ߥGY=֤MPo~hIQ@Pqr {hEWyMWȋfan.klhCjMߞP_lQAd4m7^7>Q/!, ۓjr2<Oo="xMG`^}qfdbN%)Zif J#J J# JJ㠝ZzJ$ũ/XԴuS0"9jfpK"J+"tIDnȴ>wvzB^v&#}i/# j*:J0aP[qL bCh`"W;zyp|GA/$4A WA+υt$YI4_&pK2+s|!}9~,u얪Yx9 Q؟zf"\빞Sxtϥ9G-.lV5 d%ξWy|[x{=|JnxC}7:}]NQ+ó|KƋj ϰSoyu5?/W;sK^u yMWro oKy (y'q++)B^P(Jig<0Cd~ %ΎNIGPҙW(4|4L,˔,8DK$cQ9##:0 0R@9V;H)%ZZ#(,Y*RJvmp=gs'M4hj 6Us4'1_}!hRZ8W'Y‚N¥J{B*N%LBlpQS=; C6ۻs{Wcn/픰@]DIF0E J:y-rɨpBkFE/ l$*' v,&/c|Bpp=n=WjwmQ Dre9$oܼ6Lȅ:/bSk^ǔ5IA{| 9=WmfΖKb~~oH9iե=Òݲ@j]nML}ݴ&n_m)-ePrILHR[twp !soJ/:>X5@={YB`М^XFU"]BTv b "ZB @~Dp}0SqP񷃶ڜ<po~Ufgc[m*4#1t_Wݫ'%ܘS3"oWZoG H}=Se\4m*?9\ξY=u_]G9M;:cr8k,yZ$ @pIۋwg?ZbxV/1#rj/ňRrnn-_ /j'|k!r/۴(ad|H b'{OVWɳoZ=Q̀WTj9,/wi}1͈.OOެ?xq}l%wn~z{v_Y-ŋ4QW;dy7|p٘.uzK7?Ռn&f.;82އ>hsw9Ó:ssgju v#aQ>=~1#>;%p 7coiOfw7p?o?w?|?W?zϯW/?LRer.kXԼEV\B+0ku͘:C:HXGHfIV&)nrE'R7J4N849$`7Rhkgmểen>k/|^}%&c)ZM*ͷXDj G{FJ4bB#X~D%x>Ů׺7 3/_r#|WXꭌi~oѸZ{W*o#qqȵU~?;CtbTf"fix#]_5~X ՘.JnJTL7)Z5U\#%3y9vjLZx۹ukKDbu1Į%fP&P83͜-~r*  w{.nt]- T"Y]s~Iwk {_NmUQn4lY5.P~647(֧{F=qx*Fp9Т!dK"pCS VഷdE$PBPb>@}6sn<QOmr8c5QmTӺQY~e),:[ ʃ9[ Fla~Ұl?J ~@p5 v`jK(a-}ap{ {z8'Ws9>jX)p+l! fz8p%W6+>7 W_{*! W^ֈ+mK5M 1h 3HXL :{~K/Ò_z?5f⥕oз8>9<#bR)ѳcF^(m*cu\$U]us^kwj_ŕ~ʪg簽NzG<{N^YņzNy@T +o2qwY~lF<4faR2e\VZ9XDfRMVmȑN{dC_GE_>4zïgǟ&?W(?3Sp=bLFnڳnc(!{'>:ö;wl_krl1TLqC+dԪ1+\znqlޚ̓oZÜqFk1J1#ӃI{—z(Hcgpu7l=e0)߄oS3Ljk=ն/fB ~N;7J%gPK^3bQ29S2AgpG[83dQ_:e[M1;cy#BbG&W%gsxT@ߦ˷ A5Q/Q)Uz]KQ< Ri=TB| h"/^*jbI>+i/@ɘ3H’Zq{ܜ4*VL˝w!9ӐJJOtjzmNآ?ÄF:'α_n;~Nc(aXē f͋J!eS%-CvЌHXADcd "R4ր`fYklwH!S$)!l;v4$AF U)cs Xb@xؽ:l8Up up@mqw ɬn:s,Cu %S BLtq b(yX⒲Ίqah<=YWa+3\[hejZ`0Te2xARga=gjCE(0ٻV8I%'Wʔ/@1il8`}%C ੖X|6\`@68p( NA+9XT-Hd;e0MJ F+UPX~\B24֒@* &[@ȶ2r3t+q1d a̷^vNc D`PgE,ن=0Ly^t[sv41  X5!@A. E)XJ#(yVAT6V r ֑(v_@KbUU d*3{c(s$ea#F^ߑ 638Ku1`OKFge] %Pmpp=#"/7ьV5r4&Cje"z8$9MȲBdD 贉 LJ }lkk!d8M&2Й[ [[{>~icƬsDr"p8ғcSC8Y !L&ü Z7n}K7` 똕LJҕjDRe1G]@Nic@|BBg8t:* <@"922+2!58tqXX:ɂN5~d}gڨE*ΛNp5YKaVNb;>m Ʒb] ࣸSLX.#+xmP4=JO^̪s  %_v<N3uߡk@)@Fg*D =f AnKhD;SI X ` *4ΟvA4Y9OHO,u8Cux 0@dLʀGf̱Ll:LMlM1c4(QD&@qBõ:ZoX(0lI1l̦(&?c!ud1mpdY@YIx!mƎZ[?ުi"du FcA-##*w(a5-lFP_罢5R#Kpaf, ڪ F򴱣<FE{zSq\ǟsvq T CA|R@3GOJup0N0O4Nd4^aMō֤Ѻ^yHY= 8F&h31g0^ȟL? vV1qBD r%5pLtp l3)+1qaS<LPsoެ7İh*TO[WH\M1n9s@ZrV?$JWFF\-a[+ֶO&vr 9!O%nY)Asw+6o %(LL܎C}@YKAB w KʚS|b#Z>xq #faKTYtK^.& 옏s5)nā[uႷ0Z@$Aהyo*wW8pDЪҊe+Pv&P\70m xɬS.1SPXpz8g?}>]!P٭] -\6D33ze~I!B} +{loBdiEHr@ $nIh)r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@4$[29^pe湏/N[kgB$ vY!$,=r'tg8Gh䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9_h>q@yR\ JHzx7cu!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@!r@̀OqRҝXz7F䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐B9 䀐z)zFCxq5͵zh}N&~'0h [+")ݟ㋲HbK/[i4W;h~`PQx.|/^ *KWI>ʚ|k_׆Ǣ&"/Ji!—bE\gQ*;nG=@VXЋ8[nǺZdj}J=NJU1MuŔ>4V G3f> /yV]ϳ%J3V^}޷[2_-j]vi  e7ZX Bx .ɨT $ԝɻ2,W4C"󆋫'l zfoLjN\\bh -q:oz:Ijpv:k: k,A|_q͕wuah_]|++vm=윅oS044_"^zK7mqTϫgU3aF/ES0X}*ŝ7gWp?NԓlT;=j]vGRS|751~j1~Eb G;=氪}jͦ)G8}/Ձ^]0ߔ~>n>w(=l`vv8"~`V{ݷ~:fo_C 6e壚.i-i#Zrn5%/ /&?Kd2xAx<(~7uq.R9C +Ѳ$0'4_ݼsȑd>={D$5zi8!s{q{tL[U̍.(82E{m/s s!٘ғRr%˥dqH:^7 :CZ-ⷶZ?** !B,[\jTzYimsiVS+S ,=hnˌk\Csͺqv]iaX sfhxrڴ-7][(6v>!ޜqGq Ikվ$!7?11t>fxR(h܊y&GV g[*Lj|*#۟dS;Z *WufF\8|5\yE2DwO67~x~ ?~oǯ߿{7y0Z`=L "Ã\{}=x46}ֺmFt_rnǃo/{o&gMլ1oq4_a5aj"j"f5P 6 u5׻y[}#:J;JL{&]Z ^93DybB$<I%ғ8,t^Q?wq #{[mqc(2kTsV ]*)^)-σO4D0%k;lTq6?`팟bCO}bRk;y4;s!;mBk;+=qM]θ+qM*az&0m|)?->NLB A8S42 !jLAk?>uʦ~¶)1'jY]YN9Ow$%I fJp2nj&=Z 3s:q:٥ȗx=F/y+O/[m泄B .2h|Y`LR$0eɼ!+KIm]+^mmPdD^28g6Z[žE47g֟ю:σgqtT$Vr) -5-0re>񎔩9=|vKlZp{ҿ!E}4%rf=I$ I;5 |F}r^GܟY>ptzpH7Х|(|vs+oW50VK-9- .F*]K*mrI1r=Ĺyo>qgɈ%o;JJڃtP^p +0GiKB?lHJўs۫6o9"1Q"|&&LH*c26yocT~s=)gt/hmTg{wKwWO{Νx:?<Zܤ'A9ҺN0qe>л(> G(!b@ |JR:]:G7ҕLI:iy22Ec!Q1D 5f ygKޟ.gZwۋi#=:$E}MĊG'Jj`rP/L3Kd*c>S+'RXlY]xJB%ٓLgUOAicPzrd$508 08mLEdsQ'N6)g{?KicQ("mTMIGD8 }Íc$ho7w07 z^6-:m> l $ԝ=Y@1^CtPTH'%K@11e4 cB&&p&yPsBeƏ|${NB}R.2ΚHdMx !-8H 7/ku4P{{y =v-1VUMtU2@;~h }}W}^ЁOz6:)VJcM:X 6.LRޕ&Hc鈁qEjǢmӖ[a/@v_QhS6gUN%'$MvQ9Q*N(JKTz(}:)v*oo}^/k,^ z w^W1wzRDYS%|3W{{{:u]'ZBx i("'buSQb[[}r{QYV*'IDt>m}!UނkPS™16oL3xRϕg([y!玥$kCs:dOӗ87_'ѳ_wmJ_˶؞!94P,mػAmz{Q|:ndɑ:iNu$K ѡx7p&3(TC%Ǥ T̛T\3EVef-*UMlL #(4za #J1hl)3h}M鐘S=rم3r Sv$0ĢPe+RtYEn<_5~sM0Bb.A |0&id^NZn!Q9McG4FyO=.n3lkAvm η 5$ueGA R}'W~'|o 4лuX,z_\N7kQc nPAp"XV[SULuo{@N˒-0C̥n047B3vwkMcx5V'*n W)%2sP]9kw"ޭ sZ-ITȕuSO*Ad$].&LιhE$.N!MVfP=Sh Fe&:m%`˞fL{$"灁hM;OBo[5(cQG<71 ѷmޚ8[}H$mŝ?ґS*(޴ۖ+|GhT{zɺѮR4&~a6E'jDU@N%+=Rī!r nooێѳGx?:N9[v((4>D+ڸ'=Y6JJ+ue]Ur DMǣՎmEZlWDz: H#]ˋmǴii)oM7 NeXF ^MgMRg蓿Oי>t9}M`.!͉w?ףFݤ|NR’(ddĄBT YFg;&~)9jZBuoO_x= VmzM'ϻ*N/56LnѮޫ [l`@͞z47{F/}^zx2^޽-4Qy鹨WˣS=k/>o}zQ(qNO1B:XŸe(*Th"kB&uFagn:3$93xDg3r=ǜW R`%"W6:(!$hmפR* d۔.} {hM\YX ݲnklyx-dgXo̤]KXÇ wh9SOJFf\)1&3 ɀh8-0).u'YmH*/D!Ib$JW00ߖ8'jjrތȊN>K(C&M΂[trRbʍY0jZO82TDs^3Fo08 W2̄Rf+KQ*2j4^=+1fAO Pѥ9# s.HEָJkXؚf eX(;,{yua+bQ7khaz\?)99bA0 k4F&"eNcakGqIRP(-jvHʛ=!()uO֯l],&:Y02b&fA\21[ӎ;{xѲ"`c8 Ɍ() CD}Vڸ@#r +:dkbbA$0h"Q182&fKjגq+$\3 'L -~:Pj{[,F#[;i!pnmyɂVt^">>z)^gLˆ 9"&ҞeT' NY ɢ2qۗbN-_r؃B솪Uw"rhaxՕ+Dh}եw8~Nbߐ [ L+*]뫧7T:/Z_YWH%+*9۾uI͋›+Bڢ楒a4ot}E]ݚ7T]f/akRUsYmSYEC} 1d[8k6* ܏IPp>-'ӕV^k9Ao:#}OF:.H#2(ta4M}HV@َԴ$'2֑qV'p%Y-_ 9 =OG! &s ڠ>E2Ԏ*A2kvoK-26|L=oM.fo=j,bm%IIt4 *fq4#1_}*b}8Ըћ0$kPV(ksF 4#&G~!e2ezʠLz_\NhjM-h[ T +Ap:B+ OUPl9j;iDY0_'10tR}ӓG4i{5 O1nyݩ) 68h9Oq/ <)M^_%BShZnjuPQXh@"k{u~P"'5 N%QVcOpT\lGO'h\>"q93WEڥtWG WEJ;zpev>ד~wa Ep?61ϩZ#ozA:8cǝ6f1閜L8ɄzpML?_?^끡&![ѸcAE7h)DM!o0V2K.d>MR=ޠSy%zD1v,qi'Ҙ@M>6\S`l9"EI:7uꋸJ2t$w^wUz-dէX͜/NԙgLџ럿7NO28cL`W,}v+{xO/f (MQU<@lu//?i5Ջ$zS9]dŹA(s|Mf:=thqJUg*'( W!WBH›vP_Å=.~:.*kbmc5WS5Y%8+b&:i1 j:WQAj24V 4vjKY=jH+kTQ8vP^RwAv:{E\OeH .j}{Hn H\IR;\)- •'OD>^/볩rv]թ:VN4?7zƻ=M=Q iI 1&"GA+BHV!')"ԭ[|Cv̙e\lD%jLp!{p&'mSECH֭]wV'Yx:&VתS,$9KR\tw]:w?X䜋VD. i2a`@aL1XpHXVo'XĸYd@o[5(cQG<'@8뭉ōGMёKI61h."|>^¤Mxs(ű$0čû8c̒X.,$/ t 554b^ԋTzZXh9rfl-u6il()HSFkd;3|È,)MJF!EtKrUVVTUqh]+Ju*|Dq^]MfyqZ9# !ֵUF%؇No㴧Xk~xW7EopF{܏h4=xkt1s~*?']$SקQ^43Η{SZW25ⵯm0Ti!hTgWFIߊ$ +*C%e9rv N ;4[/ _9"X P) :ZE b(C7hj!SBdrT*E0O**J`23[zzg@o y,e,򙠽3\RH&\H8YQd~-#dz^Zrx.wP C? Nз j%|ܨYX}`]=cvWQ8d2M/"E-.`Wvw!  pnxødz>&A 6hcBc҉W-Ð 5qHs1xWN^q tcQ 9B*D, SU֊=+<^OF~I~cBz!LfL`۬DnG$O/no`|3Wւՠ`GoGi莟t Kϗ}֔`uwdh1o!8Vꕏ7kqu d:R6R 1~ vok<]9gF p( qGb8SjJs&MNɅt<=GSm֧<'}K~j.~rQ g-EP YIW\xJK4a 4NS5׵q)ܾMMGkr6mIa Uw}SL]oq%|g"ѯ8/EFgXP'V1BeR.F"J0"<ꝂR߯ceI;@ߟ-Η'm~sϩx#oU#^xL|{sSQ4vݼе]/w}z<9.-ZVA{!vyc/pXwO]ܦMu1/,SW4,"Yl]u QkV{f>I;/Ɨ-o ,r*^Pjz_[ ޶LeE6'pF2@{ú/֝K4+Y"êCW}7Rx޶ `,'VdEQZ@ \(RacM4ҨzSd~y`F6*Vm;&Z"mLaA-;B[2uDVkA ״}!1.h\IJzm2TÆRq*,WD9^@W#ѯtebLNv^&_ _t6ׁh=We%6WSBD*ιXUY8gBG}SeJg_s ,k`Y,Qt"]dHd;0'bZBe3 >n~UY&u8X?`ٹ)TPK*!%,XwX:D9 '=SIqqRGM(29$ 2)4VG"d< !F%38Cm ]BD Dy ٟҎ^>qt#dR<^Ͽ(+ASP6KtFȉ@4#!,XUK7F (iٚ7*,#V9p$LߒB橏 ь#%0*'Ӛ_w)qؤZ -j7աD\ӓq5=_\=^MjϷ9[y]nMQTG-;{(ʻ:Ɠ#DSL;(Ta:p4Gzc&YNH %> o51!qz7Hu,@ &%HU(Tid,&NŸJ1,,B^ { u"3㞐dej'4`2}̩q6V2YT8ۍ@`8ټ#~+ϳnLVC #陹}mޟ@SK-; nlGbT|Unþ/]3eҪ@g@¶tzTM^xG~/ _mVbsN(twen^?NZgi=GIiJx~ne+MG'As_}ًxْ8ϏO\bϧ eFwE1I qFtD̐D 'AJ0Ilˉ& 64"\^OzgA@b@$LkS:gd N$cY)ըP@!C BmG!c9 r-QʧV֝RKeNQ APb%pIm]P$c$Jso5M'[5jkEQ2J%+-ȹDFР6x 3"Hb< T24.fQ& i(%x;)2zyi_dY I{ޫ $<.pUe\-LOw,{D>ô7oyL ƀ+IT PVEyO].'塚H\=D|5gx9E>%v2"'p&').W'pi'sά]2ǏԁhJ깸JYE0(oS>GoCTXߜkF\HQ‘}9TWD/^, 4!GN/''aRu]S8+Ȳkcr{y^g Sݮ?VTUų?ՓwW}aͅr_t|^-]rTܭ`>0x)ijiw[r}KWMͰf,Oṛwv6ГEtC2Y'ZmF%ҋNgf:8}~;\R/{O +u I*U8'YUT{?:o޽~|uۏgWG` !3$7^P4oilo47be|vM]nhXXJ@T ?_}5oS`?ikV WPlT*,?酺R_H&]|*KAa@_L{ϻz#FZ&hmbg LF UJ@J\O s0!)IR!:tpl+Jv>c\ bb9/>W6wSf1$MɄZt;L`z=:zնăRlb}wT1Nsaz&coӵ7o_| 1MhB&yh3W(J]HAjYN7M$%I@H7:PepITgLb420KmԚ"$ǭ1S)p,|vv wkl7pMI:MC™M1#1fPNj] _:{. In}GA ZhmRJiLh,Xr) V~kI씷孭c]pWݛ\d[ u3cŊ7۷y|c}|s9F,nom|9i$Vr)5z :HNK en27kf ID!sB)Ic`$jv373f%|cѥiM%Oli]Q;. *EU'T|lI<0dž-%Ǒp8Ӏv1r*HUEBDq'Ö[=/.%"֞٬Ώ8x-<bX`kF-@ lR# 8;`lqnua˵U y\vyxFր\W1K Xw-v}"슟v].ioK%IL\M>\N ru5T)ASDO~řo_e>^M.o*vƦPΌJ{3~x~~WM婝')1L|% .7ILJ[Q|f;RO=F_&Kosfz9ý˄xG`3y_X8"&fo~Wbn0^qB*$GP' 6=!nvw+vOtdiR jF*LZ0/a q2yD V[{5jR ]$2C "mΑ}, "UkO}WumW? 6?pK@Bd ONwAH/DTbbG=htRyBپh4|R(5fpJ&fZ>罥BՉYK1/[g jm~z\ǷeƆ~7whF:ZC*S6NpUdafY,Fpnޅwz8u5ilg!76 (r# UZ/!eHKX.]T7W9lt^f LoͲB,Y`ɢWixռ( {JnMs}-˜nYKկnvI8uwz!|jκ?+kֆ8i(V>?ju(oD-!?6JֱyiUmu\127Hχ }2r+zpd_>+ы xUA&$T տs}~”.O>ե{UfďC]qߞFync(©6x_̿>w< 6pӥnm)_u]US3kY,Q# #G1~6ГEtC2Y'ZmF%vӹpryVGk ݃>7w#'jo>%%mRh^_7͍إifߥ]fڽ>V-־|o-O__ $o3:':-v q^ j~U|*J~#vX.!mM~4(WFjMI {@g LF UJ@7J\O {|rP:a|;F$-O`5r`g LG{N1ҜXq6] Nl+ ~@>;G4?Un _Ʈb&IIͣn]y)M[i$G09@BQ41ĕ*+=tWz$$]]P,\s<#>IIiNџ:P-2.%+)qMI`|Fڤ&B#YtMKQXL\Nb(om{ޔ"S/۳2Xխf7Oo`k=bNBa%لrXXP^5tW|`]>c"!I=@pp@qxC(%`  !D^n>TH99̈́q~Xl#u#(]Oϫ2;Q9vh(j#8tGUrkFJ|R8g*ϭjpT&o ,ZG X\7h0Q';TkY[g`)aկ.զlT1EWt򉐥!Ȳ x213L a8 M{/A_T]E^;~M>^'>~v?5yt_{Ǖvv|?)R ^)QJ ncz8_$g=r/=`uf6k}SBRv ER"EIC'p,yT|VmKÙد>WQh*Pㅼ= u4]ַ9_IOˋ6e M}zURԝ5wϏBw6>3;5 o~,j>7\i8 b|>5Q rB#q=–kܮNo0( "qpQTKVY Ma1=ϺGžeݣu/Ӏ#v1۟*"8TH^8@Y<*&Dgn >0aE-[lV>T {@5)AabX`#kg-@@(zN!RWeHuπjfm]H)~Ն91-,}m;7=bw;Jy ͗kwzSZ/tFP\},-)D+ԛ}6ȦRNPiXb\$Um NBTAhpf-j3[lD sKʡN1%)C$ 8$H:/fpnR[l<=MhAXw0IG)nsZ(-l5HQE$˶A\WESSsDeNŔNI.(ap gV3N,,H B u5D NjqK. yt6FH 3:9,蒌T;oerw>~MDk)B >I v$`yc{c5׾b**7.́rKsK,h`[FEL1VzQ'w g+Ln*]Unkr?Kb5QPȍqUOpK 7ntXґ%RL:mb:B+UQ_*kwHMҖ:4 C@QTKo$A0R}$;fn]o-aݥ^wGQN?~o(5 \(p6)Q)FXBS\+ݶRPuCګ(.Hp_ =qbF = E4Nr}idkhi&KiU]p}jOn0[}kѶS"=y$;wQ}= 3/wwicE&Df%-v^$݉z!*ZoHtd_JԈc)e='%4;??zoŷx'XVA&PcR$o.l2!P .4X=@1ǫ"^ڔ fmc8[HgfivQpH ~jj/R{@΃7R,x@QzD "dZnȸxJkϥF{OEk&l#$$){x2h}Tl'Q#}#>&6%8EI>̪vԕ[l]3"[6/xj1 <&q5PA0GdEEEbB*J"x2;=yXg(D!Irg8 8[\80#OW@W{/9<e6hޞ3cUB& U$':oԢ|/K&)~cj+f]-GCnԭ5y0-ّgn]oqg%hFHLYdmP"f˵0:W @*>`.մ;[e-i#1|Pђ 9DQB =_a#M. ϩCUM?÷c=B3`z<]}X|&>o7l;GT`oQl\#KuO?*`;u~Y7gK$O8ͨߕ/Qr^>⬥rs%ɝgGէ?U+wmYXM%R,0(*A !S$b^yDşWXx j^]u+3U#?֏QG8lla8>MjvLat`،dkb) W5qg'n6tp<·Ek-ۆUѬmCq7u^e5 ;[?~+~\'\#-^ױǣXP/)V4ɩRR%+N䤾ƾ//~>p0Hd6=]s-s:{oKڹ^yLូyS*(SuU[v_~uuЕU/<*۵Q7=x|ȬBSbC jycO/Nc!5m`6o)';B.nQƖ{f }.ݗ*^PpCBX.`fX3AM{^:ΰM܎yh5L.Z>t^ԚV]ܚv5, VZ}Dmh,w^JZEN("A‚K6Q=o?3o#::rm+' wڀL^UrM$%V(Pi<( TX^;{2t\V{ZfhcLN6/]m:=2P.{z¥7T$'Q֏e feӘ~}GhQ8?tgaŧX-Y@e:A/&m-KbLDs:RD!{\ۆZRho7GZ{yEJ^s0tG+P"TiE&|qEY󎧹8?\YfNUW\9{f1̚{ŕ nvԞ^>q= !cOk7)*~O>פUNG|oGU/ oͱ*Y\!y3 q7O_CNDpR|"~7'鷇uM;f .c u mdӶb4!) "2*ΒŁ;0GpC䮖R %MΆ]dE%E} ,VKQc`p t3u!O!RyݶBg [f){Uma]ʝq+GR\ 7/Uks$Drٱ/\KX?e]^ m)9ΪҺlw)&gxO=l~W/)א]/ p$ͶI5Kb>O@qƓl /cl6u>pP}P7yuM>O8w'ߌ'>7p9y}-#m+o,=͐/ Xx3;_}Z:x'xKBާux,g8i,o[+x$ϧEwmhw]GśRjc+Ώ;oBw:/OF|7?_y{ЅW`sRcq1>G~D\>r祧p;W GbMnX3şlf:P tQIJB-)4g \v¯0޻\覣:5sSJ=5 (?I 4ʤD8 В1u\Q8%3xC-@rR!Db:Q"u;*J{:Fl8[f䠯'ǟpiMR"8Da 04Zƒ1hIM⟡2b 9 2KXG&ьH`U^F51:] fF:h`GWU?䩻fyM:Vuq6?#M*Fk1ro&XST&g 3|.nGJc"LGԳ4K!I,'"cۜ/䢟ʘ8e=L0Rd=II8UQ*KR]#cgùU:(X;Bc\γd&YzIw;_7(p8sE=s6<FhBQ4"ʜ2YZS>2µj N(CF6D|Ae7{D a _,꺨cU)@"A܌~< *澠vgѣv/> ܌N#.NndJc\=.iT%B2BQUDHύr@jq@{.pq_3yyB&Q)M$N ݖ eV1'P4DbK:pT.[]2Z^[N:SE :xυrT'p*RL)e{44 Ni|4:2ٛfɬV_TϘ ؉G_unc0mh`(7MW ifu PJ.T. xCx5%97}h27d/qTaEќC7Ks0\ or o$o MFi]r0 %fUn%D=w I {cÝGo#PV`$ 6""^BP֔z$Tiq@*,PU}5_R]ɕΝv)p\z0 -N2+$Wa]Q3`Z*CYSɕlx$3\= `A؁àd{j+@\m{r@*kWC!WZi\e(9WH(h%\!`UC!WZN\e(QzUX!W\-\!\e(ɑ\JrS"ʘ .6E|;\!Y"wVR^-D NY) Oz\9$W=yQk>BH.9jd{/J=u;Dggǜ$(c(Y#Q` 16O9WŚ9V]-kd->nD H ]7%#6 %U67UylK#n2lU,JrÖV^}mk6sM[B&&ca[p%k60en Q[&enh2lgZR{m9˳p #]FFf7| m7o/o[Șt/ǝ^U|8ޞr{oPBo>~׍5 'UȎ+w*>_#JbQz[mݎg;{-5І`Mh9, Q-ްr6vt %w!LRrHp U.ǭE2lo>^&C{8P^|Og-]|֙||/-[&NCIg7p(7MKQjfozt}:?8Ⱦ`oAfwoOXIThPp*0Q8ԗ WDoOxozm [@Xe}'ɶΌmw4\6N k""ZtN`SYd:Q3Gh DP.tVrt MRHngN߬v(C(B ʐ RrNÎ jc[ŻK7͐;Y/VMkq2/Wǚf7^?F{g|M:]kֈbyp.$H%HBLQ 6z <Tҡ:q䵻 ߿R/*8^OLOT)JSŽJJnyҨy! 1"ZD(1p/{(> /8)C>BV\%IHRLω!EɑӘ$AD x ?sdMOŚ 2J7`TBYEgr<,R{7މFqDճi`}zOEFM ¤"Pe!)K[B&Y H*P&9XK;C~~fi8a)o/q~L Tѳ 7g 1qDJsuvu/edߝfä{]SlG:61ga4!]]sJ@na@OFpII'v2}2"8}S>\oV!WF0 A\&ZqCz_| #l;?Cvk9^wF$$]~l*;u "?))Z{\U*6_.Wg25x#J7έ5Tci򃷗كwu.0Z*~ R[fnf 6rZ~wn`Le/.MCOci"baI>s=͜QkgpEuZ"Y$͎s >}rsX%'3jߡc82 *N7>~~}ۏ?}#zÏo w`Zւ wםBALښO4o2hsճ o2nsk潋>FоRo-/wËQ?zV7\i_A/=l.A"|k!Zd8X3ns}F}t[xDQO &\OSWBeJ^GZ521R>DY'焌=IOnpĞ?6xz" 8Gب >vhQgJcX*"[׬H(I&_w 2Jy&?3||8)t2ɱYEioz}5}#Gq;I(+ll795.r=IhڟGC{M9?*ز'~ WiQIB WXt0픅V؄S]PQ C莁zBwVJwjB(OL3 { r Ǩ-$ш>?p oݕ=iro8Jƴh@Wjw Q{eaGw\IrJTa)mbQY'!pr?b(pU[j8Y4f)mYLj bp_vКW% k#l?AëAɛĹX=ݗ/RNپel~뵶[gzNzWz/ z \*hR!T zm ܠEbJ3Gmyg>.lo񚭦4@uW_+b18[ϬbK<~iLKͅ#fcvϺ1Op+:0NB:SG 11}NRgztv1}O (=#g2j=$B i8EO +YҭʏSǂi)JgĈv!Ă2S5ZN,@zE8Ӳ8 }?cg}%(G"ƽw].ךHWЭ.|(eUM[]:|JoiDUyU:%=j~DF#_%"CV1" !K,gu4H75EP*:EKHO$TɃm[g3^t.a{OVk5:EYޣW5,Z}4ՠ(*bԷx 7A}eJtKC r5Kp/pUM4.=ꖏ[ߺ%ٿTG|/t"b$/ +h`GE V(M@9=nuKCs//r#?Y8"V$H>E? :(pJ^$<$*8ʝ7<ՠ@xcF@1b"iZ+Io-/SPyy T\l Fi50M] 7P&S=\bܻ%66n-Y_}{*eǕ)qoL@;д*l;m_ "1!|ߓ;"f^,\}8S鹝JUP3 V>eLJ9z`aReѰՇwZ :f\5qh=W/KͲՖlwayLYgS**V5],uM*4+^єG&_:zpn*]O(0⽿; @=䛓'<;5Ljgn-εkmOv\PffJ29_0IaZ1'f_i:rpxw) ~M};=LȿJk10d"yF!]J}T(C:}5cзC'WӦjKq3!gO3,A-wsi476Xy8 c`J4.ct".jC 4J![%in%aGRE%1(s*h<b>bG-Y <pOK[;_?US-;e?aXME"Z_pligxTDbTʁW&E QRYbX:$k,0Iexz;iy^hgtV!`)N ՑSaQʔa@(jL$#C\@Iϖ+:9~;InMEFLȴE2" S_SͰ~B| ǸfQФ/'"h!uB=v0G/4F d.8'Q;.D" ڇCc2v56c\Tv6g'EN{{@501\|A'@%Dcj-]R ̙i?_FKo0h&`d5T+JgǷͥ}2d4>{w_+diНaI o"k tj̛|c%E0 /znRչ™YC)GA.Ue|C4׃it3L闦كg@-`*0e V`ҡǰ'LN)sen4OtY%gߕXڇ[ek_'+ IĦyq?;/'Ԓl<͠v, T}jչYڻb?·ӻAmҪ6P 5¸\-΅ln|6}J'0yd?5Fr &-XPшBQ) C0p}oN*z߿2O'?KStDb 'H Q>rAK kOx(M?EDY",:BX K=uc+` pLRIT4``#qVD bl: Ą3[XA`Ir,i$532#b6q#y&a!ufӒmqQdEb4(H+)i8a@r[Q$5 -pqx(xM;vx[ʬYkwTnNp3y?j0徼Wgq~1W|&yBYJz Fٓѿ V~ 6ɻWth)eֺ0뛯M?9:Cy z쵲2a0&BP5;Lum`\:vFz0:wK-U3aDŽ+%*?+ֿ&;v% 6 f0EAT*f-@V[֙1;9Coe5l'#HFn$AFҘ# !ZӾ&%HAz2=sL(QmGs&Ȱ, Ƹ#aREbD h̝=8Y2 R@Җ1_+?E ao=v繨0]9 ⳽dW",E$E6bV0aQ)Px:>@u|@ V3%fGɈe 4(RFy dk-iO5{*w,0 .h ޾Lbe+yNcy~<;^G3j]m'|& c=1̏:x$ x74S>+:=,&)djO3\Ʒ; ݻ?`b_\n*EZ؞9IJ撧鴮8 -гT2) ΌR27{&_Ay S\{hM _{8WgmsH^h ; sD͍a%'`2@fYï4mY䀬arְQxe#+-q+E#"!l<$>6s wvirm{!Èef5]׫]Arh$]qeݶdZк|ԺZ^pY5v,;\{y^~6V6yj}F+k.W̛FngkGDt)V񛦫;+nZzSn'vⳄ [L[D0@^3ŕ4t&>T\D(TV;L5uFtvF:=8#HNJ$5LӠuASKR'b4'qS"0<K=s; (sH1haEJ ,hK +*\>$}G˫=Ϣ;y5YVW,7nOIS4;4u7$W%A#:}'q ;$-S~;~Gߊ*?Jbq,p*I:zp pzDpî \.{ WIJ:z1p%7z|$'핵p8q37J>dW=&Z#~Dp~5\%q8Jꃇ+L۾X }Dph U \%i8tJRҎ]FjrLpfh*UV<\%)eǮ^#\1ʵm# !\`RC@6^n&C[,z8 9$%Ѷ䈖"q"k1ϙϼ}  2o* U q%9;hmX}mpsca1q|>ƙcqd" r G_=)S#CBNTV-ά!ZQ0'Dқ.Td1*$1ܧQ|+k/Ҷ UFu:\i [++WPi;\ ƫ3opr7] FH[_DNWVx殸B֟.ˉWU2OmWM[B=\fjpLjYi+ꡦ8Oy+7BF+TkEqE.H;\ dpW( _pjq*pu⸡W  R PڛNWB*|jBx+kW U2q%9u4!1) Z# -Rg>݃i4NrA$C?R.[ |򱻳btlǒ !sxS˒.ѵuMfV/lYlMserP[XB1K\jB$"MFَe"OP*IY" Aƞɘ3BaO?5?ë+cj] )Rϟ>dF YM(DDB1Bd]̩Im<[gxY٦Z$Ì9 Nj)IgnF]݋ڥe愑 Y(bX.Ykgvr}ߤ(ܖf$:EdUh4Q&LK0a5-`żiZ@ڛTkZߴP4])6-(NĪt>KKp 5R.IŞO )@ŮE\9 ~x=5w?vI pۥmBV ҋ?7]?c *(I>U;n */yKfzkU7Sr%Vۭr D Tp 㯋C>~Tx˞\ Zͭߊ-c &:=dTWzF5Vjj .ms3 āiE Q?xVV/bZaYn} ly}J&GVx\A/e*9B մbBqp6j&WxSDWRu#OWF A|P0'B+T+[]ʶm Ji*(]\k}`H`k~pe4eʣm(9j(7jv*[#㪼CLO-z+̘W(k_pj%i;Pe˦vw: y+ ~nq*puVKb= \\ň/BGL&]eq%5x+7B+Tk[v*-NWs[Iu!f10nhclr0cNmWaj h-af8dNDmz}c?r5r}dg̜JΜȚwde/"IPH4ˉ7bP-m} l٪q]a4L(>?7(K_pj?,UW+CjX \\ͰPkIQ%NW(EGBX|_B+TYq*Mniz']5l]l(K7S+qv`EiT+X1*efw֌C#\`#ʵ\i\Q%dNWIkG2LO(= ժ Ujqe^١`O(~q2Ch7<Jv7ɱWh&9jg*epuz)#R)hz&ۿ4N~xl}B{Aiǽ;㰖+ qL?{;BֺxjEY̯ݓ{ՙXfR#1e "kSˬ1HTY$$NMV%\젚lcoH-;f*Yu8ůX@(E+TZ+TY[~ W p P *v\Jq%:^û JNQ3қ մBNW UzuP;/ fLn e<4L\MeA\u>ELKŤ?w5g—ZflKƶWݾ*>x]oF.+DSjdNtpfFmwmiQ7Fn./&x$j=, ٿLVe]\^}pf[%`mhJQ{ӻ?nLX²j:e?tYR %l.w2;뭽W.)KKAV̗T 㯋C>~Tx˞\ ZͭEK/"(MȐAZnr9#FNcVQ#5,|}9PVAoLg%j^_UjfYUUҪֻ?(vaǣy}S5 >zʍJB|i^D͋ɮy!ҧ-(3*TDq*yĕQDQ~^mc8N/ &StK8YW4x;&gp ٦Y-jyL0I eY K˜ƈD4V0hI>vsc x)2,$1/Ѡ*UTs E]_h9;eif84Ep&SYKJR11B#d"RDr2T/Vp|(CPJ̀wN_fy0JqEyFإ[hm˖B\,͛m&>(̇iئ?,S^_g#w prxyUH_iWyKO{1$M n:z½ j> ,e!w<~k<-Uߌ=40]q1aѱ.E5nZ/} :n7:Z˒𲞷'lY`.LY\עOB>EĤDN -Ar'i$RBD:TtQ)2:P%F,]RŻU1ac2-\fg|4ZkW`g=".8u2ܼ^r>x Uӗy=~??odMz=%EWB?ӛoCUy`SŕѰhW{%x2"öe6uxQd|2V??>ţ`_# 1q~'IoR >M.MkٹmC1q4*kAK$0<PU}/uy0U耩,/ ͮ3tt`k>|)&4M]0T:>JbCbp*9z ?Yp |aSjֺJe~!ĄLz(\5ٞBm]#mj]D&{e,b1{yxwL}͵/Etnۆ}S{ }kk69|\d8v:cR?5  IB;h ~2bfHc|͠S+@W qjEMkSeVH0IJ؍܊M( &OGiML[ nNHVD;Y偛BP\(a>f\\0-?z?xx $z(~3|*)~*!orj5sfA״"OÞٿPkMtmJy>6!K1 ~s]1Mz f8],~yYeO;"$yu/n 0|VQ ETkFgdjf5,`[*U>N]GA9ߪkaY4HgqC(Ji [Ecur])+\&TWZx2 G)8DWQ<wV%O',` bq/ /E{Ibw˖"S4ĎbW9zSu[Om۬S3?8w/2}dfP؆m۝iGqy!:.$~W0i|Yf%/&+送W`iodJ،DIhį9ݹLSx/?_VM9m I[%kaV=p~J& } N.ay:oh~"mb~do9#SU-qީdm 0_Km=Y$Ry?4|eVmXΨMo=2黎];ƛ-7ZRm")L D! 5)udh=ɐtL\9Cаpѱ(+j5V͑Ir+IGh\0hiĜp7<ˮB`n2CfB7!ԥ_rZtH v=b>b>EFdxUީ 贤꧵KjXaҴ"QJegcVm(,{Z1U < 0u%Yd$߂w.jfAlutʃ!NVj2B:%o7Jn%4੏pgs;>N[bt;'t l+?0n q.,dE"1i$! MJV[fRw eYd¯0{t i㹮 )n>%iqp  v9:`[`R%"3sJYZ`5IA݄EU8ΜidV o1!H  U0j3Z^,M||zs*.Kw,f7ӋM\}}QiN.ڰ?>hĶxGR3Wgv<3fE70 (cfDMz&eeȹY[޳RP[N'21;#hsHJ$fRR5c5qUj3P](*B1£ _\d!&؆>Cou\?]kl+b 2 Qr@j[К3a0 ]\pNFT7WX% C!{XIc6.5%nǬ\$!e[cI\͎AkElsBgb. ,"`%m)#1pe "YɘXtɐAp$EYET'XYVg>lI2$(SшǡQVֈrЈFZ-Z_^>5a /D@%F@i%D8KՍnqXIC8))em@D#y.zGL:i0"IsiP;85YWI/ΗɶNYmue8m3$"dF01$PԖi rR'ĜG!1/x x*8}ӇGPa;2HvY1=r5E 1E\ޏ0sا:WuPpp1 "l_O_x.KipLCqNw|1y] 㧟,=N\%p|aں7xm1vX~XVk(cD;Uzmw\Go؂Qϭ]`WG.u!.gy| F?}Kj{9DLjoDmXnގh6n..oI+˳Q5_z͜r^n%&]tq6%v Yc $6ڸD = ݴAD *|q6ʨ_G_JofCCZͯyʲ:eŏa5L\r_QHa,#Dc,qBr;Q-"Lf3sM}7VHsiե{ӫ.,[ϥvSjBj&et\k㞫9Qrp7L >E%Eѻ6֘5 '>ކmvG)-2f1k$p(7'HbC(`-GbXl7$CjHƒ}Z~y\agogitg [h|Q~]hA~xM?vaߎFM9M{7;4l8:++`}@yW>5=P+`݋ڑcenz&Her ˏi|!WԁZY#D17fQGHzs (k@Y;Qi7+>^ZlLH4ՑVH~NqY\߾8{9vzr¸<]⦕Kzy{x>}orM/ǁO}z?.}L^WÀW :RtQʼ1K5rm޳mvG. Uy9}_) ';Kbߐ l nnۻwtzf;[vH-jLۆOG+-χŖW{j=>s~"g0.ґ<4Mw]٧/?|-s3qs.@c8#|;x+pLEMAh(6ΛholTeâ)bHR$ yX>XK $WzuѲ=85ră!ڹMgW_?vM_o7]7Ba| ^.κ6+8)MأX $ZT* -B)t<fHB*L+88P[N'C@HoXt' ND E]8WZ qyb1y,Y \}fyq5Y+Iq3#M#]4dhJ߸F:! Ȁ&fϩ ɳr<-4c$b\Pc FgU3DFDv)D`= sKs_p{5yY΋=T6jI[ 1&BKkN ]ko9+|]l7 {|ښȒ#\KIHJ-yxyy/u6(:yQFB(B:FVNJLh${ lrb9mi^{ypazZsI8BeỌ2( sp~@^!0BOɅ'v8W{w2!^Į)'צ?E˱g%ojDg#1 qL 537aty:ڻ8Ͻ(F|q:`-zZz)׶I+a˾oR^D(J-DЉ.rZG؏= ~2|Nb&wZu|]ur"9{h̘+Km v7*٥{o8!jZx5qub~Q؂Lu =ٹ6<%|mluj] ?0/ >?>}3DJE^ ~ 'P{p U6*O>._.|_~ξCۻ?|8û8[`}N#d~}@`GTV4ߥjQSO|zuއ9Ugsk5X|m/|wwֿ)p]25cPWlWZ;~RDeՏB k3 huix~ȻPk젨7j)$h챃B'Qp+!LɔGO",QUsBFp>d[ S+g؞il}#P24}aXR zS5σO"2T}Zuqts61jKk dǕ=9 |\)p,ߋĔf$;aoMdT PkNO,|g'i0؞g.e^o[}R)$,yn1"GQ"@#;M:A«7'-o^OᇛM~W c ffؾӬ־*4+B:8qQY& Cr-gdLJϼ:ˣWM~c^t. 碭Vzڪ:uYRw܆gȳfnf\† ͟x?ji~:KӛV0}Qӫuu3c= ?)fc[\'{hXo3Sd? OWَ H_e^2p8szYz \=JEWpZՃU6Xl:Bk!Ml%p-:"S}4peGWZWJZI͍9"BGWgЮ5U˦URA ?"BU6ȱUKOJ7WCWȄX,oMOr99;uZ-V#H "N;%;G[f_;P-@Ru 4c-[jM<>C^QgJ#:H{^TGxD#-9#4JkMr+Zϧx~z mgtY*htz޿Z6!s]ϡ1Fnyv ^/nը^z*Ҥ`p Dِb$ÕL\C0yVVRZj Csq8Q$w-#w8HeH 4r-Z9j ii% >kLT^QYӣ#hpWj ۊzG;Q Z[ږn8kBf<~4yDxGYr(0HCUQ%!xb:FH JV+2JV$Cb@z^ q4tJ" &6s#c}\6,g싅f,d- ggdf&bh{\Џ?nr,xJ !HFD$eV֤ 0e#Ce2%U":Z;GZ Y\CĠ6P*jP뢶#$MF&)]3bfvv^XMAzG-= M(ԫ8sD[@8 :"HpȈ %Qz HQG$8Mg\p^(Ţ4YSR9;ҽⰜ &uKEY3.[\ܘ @sB EUN1,7 ʑ 1-fԧ("12nrCbS< lp9rjQl76B*5c7q'MN}(黹A`UW^?U nH%EʉrvΒsKJ[Je"-5Ӻ=Ν{\hC(CY_^_{@>G./wHrX3y^pL<&څhB&FRˑgjA"=&b}Hg伩w04`JfM;r.c7|]~u TG;(YjHɝKK.bZRTTTTFT砝ZzJ$ũ s,jZЛ {q- "BK$YϬl wqy<+^G3z|(&5f!a#4b8Ԗ\%S:P,.ȫ-F 5^Kz ɠQ-4^<ґd%4f~HBi.IH A;WX_׷*34,{9΍hڰ?Qm.:IO SxWﮞ?]\<6go=J6fU Y@yC0~~k)@*Ҁh᧽~ftci9pb*< FT~ os +Ora\:na~$XD0U8}ƹ*8! ;"GZqNy8r1\xBE{ :m`3\I 7YUI&RuopUpS\JԳ3r[_%NNnf23Ď'Qة9ӫ)\,^mOQW\d15:5#iNub<*LƗE}ղqvc(g b.A/fs.Hp N*Wn> ^ BFZz`h8|hn.C6ߺ]5mrøwǪ[W9Эq~Z|7A7Dжț-Zsy b 旕w:BEE?֨Bjp+bІx!ЄCuޅwݣ#Z'у=)II&} ZJ*`@J\O bB$QNR!:?i++Cl^A9C>tyq:$:2Gל6ǐh$~k봲3`%&GxsY;MƬY;̪.uyRw@W>ƫ} H|4$ؠu&4diH͂%(h6Ďy[az;$%xoS7/&ķ:a/+vE~}7(AeG.]@Qs-,Q܆Xɥ(3NFYJu\Lw¶`KEB&Q{<18s$PJl |^{S0XOVciLjjԻ˖l-M3b馓Y%pF Um.D@Ϗ*`95{J|R8=6TR**ku83`Qng*`&Z3';dk[#~hg}|RSG_6> ٯO> "sc.-I:Xf8M3g#BkkU9^K,h`SFEL1Vz'wCe5\ uUz^zd {U2.HW9vXG/ Ÿ:B51anU1TζK#=4AH[Ƿ꼍=h 9ZESIbK/-YIHReD<܄Z/03Ż䢘*U/ƅu 8r֖}?6L4Jhi .Ct2b !"tJTģP%FM-Jބ^?]@qz+)Ż lu; *PѣX}'}zhϾσAJifK`2ޢޱа\ob46+1o fSY'Ad8rU~rbI/EfKe o ߙzd1j2`1smt*#Pu@w$ ,zYt);J)R?, :~o?7x5n~ءJgu@Սj\^kZ:H ^qx& <p.I'y-˸ `xVZ{.7{*^3a#$&Mq`Q'+XKoGIi8Y'%.g\kueف-wS4Pwz `ς$44HLȵHcqZ[P&yJB%]mp4i=$uFL@8^q$wYsa@ui @?-N,9>zpm)=gƪ+0tD8VP ,7q1T \-GQb)h`D7R6uB1/LJ-CiKl 1e bm"K~2:W* @*4UӦ;VӦ'tKc1yr %GhJ4LC&' c-_0)'\ӣH#|5C.zw0qZ|5 ,HCg7dtO^VBP Go{8~2Lݠ2 V}@Zѿf 2o|vk'z#Xpol!czpFˢ'gHs3*_FHnڟT>%o心;y?o˘|*^"+ KPjeDPvE?jN ^34\͟?uN[]9W|3{ Bg-{ 4WcA?3~7//zv~M.1o5{^Guև=jS{ߜ~rްj3o c2=~mWfqΏ?3X՞ijvp-%KjJ/+MrTTI.@ Gu\rRo^cxsfZszgC6f'.;M˖d c@[rHne:op^$m7?1MLͯ'&v SgPTc>W<ݛ֊s"aV/=a=fe<[&f\u]޾2ŮncDq"*|F{HrTA8)U4"(-D.,HT8xm wV䑟^&OKY-lzO:mz\.Uii\IJm2TcC8gj{9W)N$]t^\"`lb?=LOWtUWDyѝkYY-^hO~tksci lx4ι˫j'5-½U/D g:"B/QR(K+"RJ" "ҖS2* z)(動dU5ӻt{)G;إȗYkg`d3Y }޵>f)lh4}AOvA 7't!PdVpx5B#R8lܾ:6u;.AI:CGR];I2vMekvR<؛K4μl'~rBQjU{*+o˴&`=ꃚȍϳ>ÅKk9L ;W*ǟ}l>&b}mƦlge}8U }R\@Zd™Ui4'ҕ>QbGZHEtI>={qBRqH|5^ 聴`^f{Q8V&/OS[{]nK\J8H3MmT%20GzM@Ozvk<\П yޫY3.sԫٹz+8k̻[_~eC1Phb^QOF#'f RdKO!\2+it/ȉd{/#M* RςF!')bd|E4A`>)4~Uxb<B>罥Behuya4y'&~9?ηH>r8XHD ^St65ݟ޺itMR]|:}1rbg͙WpYMm\ye b++K"Õ10MrKT ]HA ].(H{5L,"A(JK4Y Z™1Rvt䘍Ԙ`3ʀĀ謏 +ʬxj9HoMMHR+O.@,G׃՘cźx㧆: :La2(MALW r#.;?]]g]R%-Vd#G|g%Iӫ@tݥ6V^vY`i`N&ɉN.o > RIxT&R HiA0ʕDl=E^fm)ߣfQ?Y[ҌOzg|9p GBdP{E<$ZL !8)0VQ>piԚ:@,H'9 ;GL*VcH-hC5q6̂Yӽo3U¹,I/R\ۛQhߝ^l&JW>zz ,=ݠә=r# wBLRD}(8r:FQ̐AhPZy$H &6+:~Pvŋ`|M0)mH N$c1hFSQ#"H5@ p}>L#W*|neݨTF&5_ t)&P=MhrD'#Q|n?Gs;TiEF%(XN3r hǍ2A!%pKsֆeEچ-*KYICiK`o0roO &P !8B,? ݅MN_cef]7+γaB^,irwQߞj_7Յo'Zgw1r6} Rwv^-]X 6&٫s7+Rap"iji۫- yKMͰfmlfqEE0V6jXgɞ|8NzVm߄9k[+Ǜ_8d)]k7[^N} ӯب5\o\#*] !{p+bClZQ'ItoI>8AKiZ dnhgUQ>Ʌ"10 ^*>$\'=䕕|%=DL,o6`s.Թ 1yuiH2ݾN+:{ڜ؊~yzw>w٦rcJs=KyiVR|7o!!Ti2@pPz\(a+;ŝS/ӧ_O@RJe9*433ⓔ$9H!10,j '9A8~&Ԃ-VĚΌ ý7_>u]|OX[c c6)L&KCPYDQXDA& 'SVwIw+^T P oyxic%[lƷ]|L2"r9Fg|>I8ۈp +e^CkE:Ўt_@U$$iP3H"o$ÐOk/xj yTJpCt$NrT*g8P 8peؚ8 %n?zyGW޻w]u6"oC,y}Fe}??]h(j#8 7F F~T˙9T1yrU wVTYv8d FĽ+?n0N:s!F98jfymMrB΢6[`I}ʼWG6YisqOU/&[CTfqK>WjB礂iV3QxQoea''pP<9-'?'oM¡G?ptG!h7#M\Ip|nGܟU[LV ]Bg^gŸ I܆|>#\e}:u9Vl>YlX2!@,!Vab:.~?}R(Oܿ9 L_jqU]\ߌcū/jS3,q:,p6ÛLԜ^˧'`՛%uz4jJi|ŭRʳvHuƔBYz0JHa$hNYe?o:8eՆ$jBDaԛ@y::Om nwwCi GcBM܃  ڨ<QJ#0o9mM r UܫHexhs!a`"#Pn(t]0L5GCG&>LQGe!K#= -:k%-"TKo$A0t(iDKmVf0c\]7yMqƉ{##5$uOۥ_v8 @ -eNf5,mPSR$*ɭ5>p(r}Ulwr0z\Xg^*#`Pdhm="eBR<\hFd#rJ ʷ[GGWR<]r^T*2(yFJ=%k46xxCg"<Ɲa 3s)S &6B! a+cc&Y rK9~bcg:4HL^8طC[PEIG&ؒcgzMOgy>7cQ":xC1*29n΢q.8[\$0Y*yؙ;-Nj -9qfJ8 Ð? N[nOǒIJ;ўa9=RJYm#GEǻt6L67"YLt#[^INk_eˎ[XJd$nɪUdj!}(EZGԽ-QXO ,"{җ"m/_A7$ƳѲXd`(HL$C0\TF;3S[6s>ĝ*>!zd@C&@ikKhIobWq^rkzj_Fif0 doz֕o?uAi8 Q#_İ+bHP_ '>ioN|J3*נ=#|{6+6i,?zNzţ[u/B^@y[uh^]5È~^`\yl6TH|<M?g[/QO8}+P[rUU$JB%I5kfQ8xrڠAV?3q^>#-%Z3nKq\Z<ײ3}.wޟRbx=hբ`]f߼XM3uU߆.s7^6aE٦^񠹫5RP 2dTiPG#Ƞ? d: {^+^.ߙK'ݥM'r{r *g7Hgc`|4;Epx7Ɏ?qֿ͛_-3o\S #<Q@^zWr14W\tc= OB-q~a<4,0|+L"J?.ó͍tXr|Nt˿M~9ۤn_%?Sgڸwt|)T>&c|/^ .-Hz++Cjb&)|StgI.^= ]ՠBJez0j70 CTq3T'8_2cyE:$]BJI*os4z@/0f8;ꤰeܐu0%R4&/Ie {,a{=vGb ø6IsOosu;*(sh˞lϜWړG^h>`t4>e dɪ$^Z9Ȥώk*KL=^/W80U_@m%7޸6^ ^> Qٓz[^bHm={U:jE ǠCvk#myl>(u)^R%OU4,,0Ʒ0 rJs O]=JJce+U5j#aMYH#f C"9Б;&&H+ΙCJߌ,DjJ+TK%TNΑ;s2A0 wѾY.RMƣxĮ %H-M44v]^,IO*eex}|x&kY!*@NEyIeOh6ZiI92o0KaZDhUVRr@$,H4:ph@PhoI5,6ɳ۹,+a/U쏤L,7Al6k5DQ᭐ +FX$/$M|CU6. u X@[h9X [V$\K,`bN8 - ~Q\N(c#c$72kXt 9$Z4"1ɂd)I@zR9m(c{$cSea~_5?U7}\Ҷ*Šr#-JY?.U+ZeRQUٸVc~^t6r9zh!dfYcf23⳯ C;3>./4z#ʩR=J<3T*8H^*QgVI)CCUqP*MFzS_خuԲ6vG5OLZl#iZm{״NtyQAjfyp43s׃e48U{V)u崤"\.4G-DeI푓Iaq]F4*fBHIeN*t6*qkD•='ht.Nq2!xrYJf.sx ˧7r6)] t^zοSVт`|N3(|bRFdL$䙌YgB@̛ I@uvk^%yDHH8M)X\YP{#gbinJj)NKl.Uo|r'NӋz+ <|O atʵ"7էf 8C+*o1Ǵ1}  p-SL``6,fa18X wdKYrAЃEm>\ H2H\+R}#colߛb  w*hGvudn{Ygݝ`<|*zrv2Ex&1<GRPVj~äH'n(!()M@֯l]Y(ʛ1h]ψ9;LNӒ/wQfY])G('9,#3ָXAWA=kc!X򱗀V\%Zš,EC#a]N,Qg<썜xXA{mQꈈGD\>9Ƹ%{zE<S p@##@J w}dVl,* ΄Shp+HPV*=#boG:v6">u%)2,}`,LiH)G*3ई  ꈋ}޸c_l;5v5 8_FQ;=nw~|,Gޕbl :InJLsR Xu+c=rnq{jWNvO#=.S!Tx5 䏋FّJGIUYp%w!!dRTd.HtKD2>4y;`J2[^[[0P/b(tjRr|\rJd"%/){\0/ed]倉hR\ꃏFg=;n`p%J=0d8΍2adzJgkbXȎdBln4 'Oȋ'ay(1^<` ~5!auq>fQ+n]R/w9EMa(+ FZ9W݋Z}2l'O[ŕ.*+](sߑp0LR%mћz{wf¢?4|-Oj<ݻ>>MP -]k8x\~QϘzhQbwR=3?R4iΛL7],>6 n\LΧWbpy:hgzЫ t'W"k?.\ݏ\)GW\{:նCϝa\ك+"W((pEҮEbW \ G?3ԕઐNBWJԮEr`!G*j~(pU5{o *<շW5*A+.Cw@jw¥Lkti4mԃN{ @YN_[W@9Pa^6#...Sd=#DRA0rMꎹ9wo-RCp׼n@w'#R:Te8999?%r: VʚqάJ"k$4)Jk =$$v:$4Y>u4F(rN 10c5HXќYy%;&d1vCsɤ@3$n c1uh0VcHJaf,EPilG?!n`KU * "VjVa ,3*D68+  8csRȠE $;.jՀB+"X`܀AS@|k)4VH(,@6{؍ttg-JC(]e֜Gk,,u0wD )_J&:&|K@hA:2]:∬ @M 4ݙ@Hq{'gGVcWGE]%A/[.ѱ=#i̼BU3_Mr^KYBD jLc`V.2@b0۽@VQ=r+`"Кўi? x\XvӋ~)Ƭ䤁1&bb٧NR!&D̿h!ìrL|0s~-" \s *ՂYe ئG0BP88@{Eq:[DҕjI#J2a1'8 'ZeF5:/3(x$ H&rZ5ȼ`>xmBV58 eX1Fz@^BB$>yt_V 5ȤuՅ:vdl>:d,XN5~d}%y E*(8m`^֊N7VE`ӍM//jO]@,M$C.2The1+pm i,{Х)GpC$*Kt]_>1&!Fu05K p[O hg]+y 9.z-| bhw@[vThň޲N+}Ƃ0c~6 3jHQ(ͮJPƁgUkUƂy0* !.}p$,/dlA'J sa0VQ{$V̬$޲[PRS[ q/G^c-A6@*tA6 |+%LalmRj6Xqns6e~1ߗaL;N&:b&QU#[.n[f= ]Zp[ LSv%[Y:6tk*ZSR@ΓDk7vS^_&Ѱ[ 3bO*V&%<%2`+9P.O3Uw-HףNJdt AP`m'Q%fdiOdPH;@qX 7n#2l)b*'k/;lE^QH"Ndvikn/3"7/V0 .LʴBѢa6 Ed$ӽİwuchp?{mDVcS380)`c37  5&EHI[dݼLFLF:uja3\Bv ZuZ"Ϗ*U^/wJf֫YWqڂi#+(|u* FWP '=ʤh]H,O30%7aFD{9>X2'Q=n!X7ϦrшUDL0r( ;ff5ФU\FCt@IWKlI҈5F,?u 358 N}`(*2&}-&h;`(U /R~? (G|1nnb] Bzk}[4]V%h6ݟM}v~fMrNjZyx?);5ݐd49hТxI')|[*3+ADqL{oV 溣{ ˭h}V_*_h[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr'[EHwLr#xNşJV>k[iXnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Nnev2܊ sȭ`[VdӭJ5,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊VO( wLr+ȭ`G#QxtAH[}r+E2,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V#ʭ>/(}zO襧^hڮaZ ' 9կ>;MWT<d0 m]PS__\.T,d./ՔWp%XId:ťߜn9ԋfqQgy+o4{>a_apKe6Tpeesgz.L >hXN!lL/M}ޜ>?ػ1z21j` WC~pICCrp_.!Ikv2{quVqVxQ~w)Š΃I 1>~-Z11#w7}> lFf>|hlsFe£FY ãQO @AcBWfp>6cwq"ӆ}g3ȱz+KΘM F:DoPBP6ZxG NKZ^-hބڽW dj+|O?pG\%\OpEVZ+ WAHor].~q1zrfsrK|0v;C3Nd>˯s!X.t PLqQ D1X"UlLNֆ-xؼ` /߻CN @nwxъ&rk^Ŵث/ 8RX4ٖr|9IWU*?~'ӫKWg0EE gXi r_ϧ͜}z--z9s"(.6MhmiU`-YGƗl׶` m)_HFUi0!KP;[4V[N![t~?Pgm ˋ?[{'[eru_onlf3lOݷ܏Ge|}ח챏Kr3nmnlv/볳Je 6HU5B''jY LܝҢZv[n[cv C~A2EY%C+]obpXη7\Ns~w- 0|픻.}=`ަ3WӋN.|1I?v'-O ;GVo~7u[/fh]ʯw5ozzuhKWȃͧm&MSoJ['\=ͣ;';\ȇ{s p7{{?iy{ݍ^ ܣ+l:=mv.oLUw4ZVZA?n+Mr/=,n^s{h[:E}RiuRٓŀ|>oY\#z%]x%{%V˫vՁ߬ھ:_%&z}}^yj?amč/!hI!h/~OP'|W{J‡IVnp8%1sc'h܍>Bt9<oϟ5oVy4ȷ|$KA:截kaVS]K꣩%k݋g WY]katV'r=3W#0һ!x-mr 9TffO ,zE" ٚ@z۽Q4/u jqXM2ڞz#*{BZ%V?$Q"jb禇%s^xv&`F1.l1ƠauktyciIY48)ǿh|>jY@>ײ|elt:dF-PBt:Z6!*nx[N; c ,Mk5uQeݐLPx{7V+gҚV 4dd*c#̹a!&5ÞzɇqP{rq?wύv,;'/N_5f[, mTN1XJtw&zP۵ZTLz{IsW<f8w/s@K^2B_#GfĶW?g1NV[3,^!f}ڋgz8_ew +g8 OCK5E*zD#QP1`QiNt*Lg;X׃*4C(͕88\ 8G!d|0.D>Wl?QE9[{rOO$ٸ眈]lͶ;jiV؆Xns >$<@CeUơKq4z gPSOGfNw,l]ZG XzD$pTkDqgrhZȹY7' l}jX!W:vݱ>E}λٯjzu Ҋ>e}4~*$Si!%6RZDf&hlSwők/֋S{ũvߍ"wr,Α4^h%̻eVB@72T„sF[ŤDA"GLqolg<"dP Gh:@7&m-˥6(+w+/E:3%}G:6-oo1xsdͷn~xMôn6Y>r?0b $ʛU8r&q IW8D 'AJ0ItM1v~Y3ۋuz\)mA'mp"+dFSQ" 2Kfp/\i]P_Λ {,\1U8/\H棄#by&./TiBˋBaV^M!+|11C[qJU#b歋yUd~28J1%1SzlKdȇLJ'n _Bz= ӦnXs76vsYޢ F*#G1[ppEcD4.^׆w(D]'9 KdUW*r$9=Ũ^k'E蟝 ;7o=|ˣ_oQfyxK:B@pH7 ?ߋJA;]]s#Zߦ_+7{U|}CKZ f^$yj)!/a,+H6il?)UZ !|4V* L*^ΣGZE;x~3I*` 7pV%.'&Db,T@}H>OzB啕wCli5["&j8J \x]sxCIM[ie]g=ٷʃokh'Vb dHa^x+|}Vw|;rQq;K<`UuqX} 46ROS{yΘR"Kƕ^ ))Μt7c8eU Ն$JB4Z$&PS($yVqzqc( Dh,@{Ђo[1 B2*ZF-٢zyqjUݕEe>>q9ߐ`~Ƥ;?Grqc#h pTF}JFz4AH[$ rPIbK/-ӡ$^(wPK1)>'lX9;}[x5JOp`́ZZD!Y ("LDH<a U[k4OvSJ r) ܕ>@qQ VZC1e+'0,.̏h"TX6r끐pqɴnFW9nd]|0GןƦ:w J5Cfo3b_`~ry%K_լrěe@=K9*͇ʿRQR?- :f J=X eYycЏ`ѫ}y<>ɺ5[;}}.hvk*E''C).QMT#W{Ht`#xBxo8t4N4B9wS xJkϥF{OEk&l\Θh!oT"bwZ`5w`4$yJ 'e5?f"[;o=13+Bq sD& Q}^84C]PEIWX%3OObzUOy(cD!IrgQ8 D h>j4O[I;ז;tEq<dK}PhNqfJф˟D'Q-Z' do'{iiUh 6CC7u+PԶg7 ߂ȖRJ< Gr2 `xR)hj9Ki.'F$2 ѝ ռ0A(iLvr2yZu$ƚW.ִ /U\q!W3s>#aip cíwի~JZ o20^  ?ĭ)3̙J_>֍~mG{I?V֩qF˽x:oƽ|{<-/g-LwrFM#zS`zc1"8xK(WY CD6P3|̙1;MyrR]&0Y q=iz3!= JѤ};Ϊ y[ZD}R\Z-EaNeEe8qI ϼsTOϷ%&ƼXE51(Q.%Z#Kq+b,ۂ{9q?Y'iki)||mfۛ `WR*~ *˫,vd}͓h_5v7?Ga70 z果ԜfmٰΨ}_30';&a^| Es7\=yӱpr3WM pmILмIB#WpganfolG 0lI&fR="w~MG0¼vC< 6*]QhG$DcRΆVdp RaM(hO,?'{[I>J--ٵ+P7#XeO,l #ߟu/j17,R墢L^LԗU6b%|W8j|98 , N4~YxoΛb涵][rXYXqwa}ZTa.ޫWX_2lEOUEEZ-l<ʷ?~_[gѾ"+zbyLAXf%gџϳwza9Q;~^.Qz|&ɦPIg$ԒRH+K#9*tw) chiivPJ=5 29~.siF_阋W$ F氇#+l!)gFwgSQ]BA" W ڵ>45rɉO MS4wF%Bl`ʺ*2 Vu --h Pj?pxR) Hv;P(|lʕ(2*u4J:[PGe*ĀӞGZE_ƪq-o_uS ƖI6_ Aa5S8ah 2ߞBmuɹ-c˺)Z ڷׅ8MX+_),V{ [4VGRGQu[z8ihw(H(h|xUi,u6ECC:jMF]jWVѠ(U[65 .+_*?l9tžF2`d hf!2{\q[oz{q~N0rzQ޸0T(`s;wt5̤q61]YWV=x+`q5JuŔh+O`8`q]1-uŔ1GWKԕ1+#FW]1uŔ>f]-PW6@ HW+ QlS(Tu zo]Wuz㔎ڡ5Qnb/H=> 澺,NQUqڳU֤ ϫfŒ{+.u՛* 3$>ÇRIsSsGGWѕֶuq?`)齔!w^?6(]~9F~us뜣7JN%v-Wj ҕ~wuJ%:m- X7`۵T\TehpZvk]fMmZcʵ-_Ǣn)hJ[Y4:>5;`rDJ eL-0451>@{%N-E/ic FW̄2I^WLi1j ,]1.V',)MȺZ6+F97)"Z_ÔW-QWX+r0.UD+ujtV}(4ϟhp;(ܠfG;`pIl=UȺڵuJ81"\FVcbJ+/HW ƗNW+E'EWDI]W9rB ]pD9ʉy@ɿe 芀#D1b\-fΎi:Oj˃ף+XQʫr;7?qs2&<G vz Y+3Q5J֚uŔEWYW/+ J%Wr0]1nRtE)uue@+/)"`VW{'EWDk\"Jڛѕ+vF)bڐstD]9pAu+Ƶb殘օuŔ+]-KJ`J##p7G۹ƕ%x<sRјfJOHسq r $?wOW9\N UZyRß3B FԂTJL(゘PɯcNVuŔg]-PWZGQN]1zL]WLIu](Ab`''bhhJ>bJue銁m+uQ6uŔ@ͺz]9~٤b`qQT;Di +R@<ـ='#zlr{z@$M{iFG)&Zm4SƘ5@MƔ銁+E1svLm"J jB HZ^7x)b(M"uQ`މL+b+몣̃W+UQE?zn]v}03IJ!np]]}6jqtE_5]11);.et&{-FW@SFȺZyH`kq '+̃E:mL+v/[j*]WNv(ƬYUtc͑G.ccǔ8QNiiuJ6W Lejy$~S{M 'k&@rdž-ƕ %xR8xLR>f\u^\bWUd!xxwgAClF9ClƍbDU/n{Cl۪6H}h@j$+w7/o#k~on(_?WjGGSMEwj}uQo ߺC/zֻoj…n|;hp>h׿[}vcGj o޾y/\wǗ䁦ާHmNg2͏VQZ&j*\if)xw⸡k5Y%'_'o07?zDYt[P__*wt/+mVh6۟ίqngצj=㤫~6l%=Q$j ,(Vc(fM RfzzpJВ]1RtŴ372]"GFz1b\giSujt6[u58νuK<'Lm 3d]ZsF8D1b\D)"ZTbg]YAb`tbtŸ3N56:_WL 6j2HϾ~gGWe .QW6@]ubLC"JŀYWѕں{XSƀycG;(0p',ϼE~7X$!c`/'` fPĴ>s()cr^b7銁UiI]WLu@]VK뽗+''LTWLyyBTFI9b\iM]WL %*"DB?#>jiC"J}իѕXzh,qv]6fӊ]u3ML-#tvz@;7+btF)bZҨJ# ]1btŸIL]WL!j2T=\|0GOyZ{Tu- \BuSWjPxVt|zv=/{qITzeѬ=>:ڨbL|a^ran5'_FMrwr3ӥ>RiD ^K 8Qtit/dzϧz#΄9=Ɏ9faΒ j.n?ooK_Ig'3 s,X̢Kguq~.RןO t>9dwO))pkw>K}S]/ߴuwJMwީ>'w-Cqvvujiސ^U 'jEgo''ِd֏|PMƆ|ю멣CGoxİ.7n~ZnP P{Wx6A#u~]ɣW_8ߥ{{";c~w;1ZSLap+Ι)Au<׮1wǠv>|ǜzoQoS=Aqv]WŦzZlbI <0QA~O?vI ޛLSp>J?o΋ 9mq7]q 1G|?y _؅t0$WܳNSy>b m Ь2FR>iכ|5/tآMx%jQs*܇0z&`nSg>sut6תvs.mr{80h͡HW]d[p{ZutX#x0C1׋Y2 \ଁk֠5G:l!Ѽ#MQ]/qaZ>grݷ ."G6xs@g[o<~f-fzfzwq6 6^Z\;쐮w[ kǵ-[] oǒ+` dȾWa y1b{X4L V8E$fwuw_uWw\H';MZ}3 ay4Y"!D-e%*׮}h| PJJjq, HJTbը'Ӎ*`v<@%r\򝨔QWOP])-G`ɏG]%r=u *֘ӥ#+HPZ]]%*Uv+FHǿ VK؉nI#_y;j#e#%-mվSI4="uS"F]%r:ustwWWJuETa}D*LѨD.;uURF]=Aug m|]*D-a7 +$U"Xr9bǢU4I+qρ<:}/P>_7Eu ^9+ ~֟Os/}UTE`e5nˇvJn rkN%K*:u((j1 si)K"{^=3X1 3 r&]C ap) ZI8F.1Rʃp-1Kנ6Y#"K9>_JR, Ԧ/יCMJ]Fxfɪ⫳KϦ7ꕯ%]AuB Vdyie=nr0)0ɲ#xxMnDBէW7W}a]bZ6j430:*O^'s!iQ #}"R&8X.!e΃ s`5^Dx^^5/X^G*#R"%^"9̰4N+!ʓH&ȍxV }֛a +la݅aTLFARyǰq>( d#!Y5(<4=iEQF ɥ8* P%vJ #$ȨNFjKf7Ρs$(&&b1Y~ãBlBGcDqU bIK%)a zub;r˲p?dC?o4 Ԣ2w4Ù\b fPZt_ҶDi㳁ޛ!/Ɵ5t.p[JL.0"~ Pc?Q}:fkQv1`c{Ү1-sX> )=6w 1c9)啳9&DB>!h[ɝ&a|=!"}k)RR+S0Ն,s&h)&F0 txxH/wۑSǠj'7ŠK\FnjZ)shOڗ3kisV9s{1`]'7IE~gSmG+xNJ﵎BH5Qc8$&z;TFxk9z;$Ex{S,@zַeP28g­6#70jo΍낍Kr}<3qRXN9W8g\[raT"k<8q=}{=׈qh(68F81*1:#5iVH r:,DDeʀ#"&ZH0<)N׌k#gǽ5.'o =.D\vEm6;w>7i, vpdU*Ze!0JMQ@щS)];y-Xhm 2:ڀ-b0XJDPB25Kpm,[q7 )3x rX;Z'[ §( [~ F%L> @1X@im,0p[(`eqasT ^LeUM'.sY u7]wJ"zoQrlqCwX nX̵Q*g er+gI1i&q>1kA`\DƸq%cJpʰvP 11T)b+t@HmojRz\Tkb5ȹ]VHߚ "ɡ`a[ġbq ޸3Y?FB6o4 :zt$>%rpP%x)B]"i 3&ه޴C'NC.\܃}`[KO~8f`ܴ(wt3jg>LLj9+jphWmTn;{6?PQ)GN*0)Q293InX[8O9K꥽~߯z_gi ߱XkAi&Jx`%Q{0Q&[ᓄ!M<}Fvm~6 7tQzvQzy?<ppya \mCK*΃eYCV,u^k4½hs6nCA1(rn3i8lH$fwNY%Ǡ5BGh[KzhSBҔ^Lɻni꠺;̣+(c9~`ţ"%k@2(b2x.,&H)&* kR<SzimP*d3"ʼn:rj4L"68SA%`~w""tEqYs;InMEF\ȴE2" fX>!s1nF}RH a݄2+8$0tQEu{v={3P0- /աē ]^䓡{߸":RhZ$&P7;ttg5ϔc+#)=>sj ޝRe9T P}*zL{N|6N̙vu^{N) ޼yS|JABZ 17R@f03͓RR4GYy'_) dv>ꜗ^=uLU?:fimҁ' _Ǘqg4O'ḓ򎍧/?_)yX/pH~]FiV ֟0>+f~!" $7)#yXԶ<{ T} ^wPa"8r}.fFE%h\u@pkm#GpE_vf`6I߯zڱ[q3H)v]W'^!͕` w2YN*ԦXK7 ga;Bu"ŖiU,ZyŠHTdz+LZ=U#QY~4/P{6~Dc'DcKjRWo/#ߎqUS7$_rnѱ K'=q[/dvK*.>,ҲMu3ޫۼLe{&fF",m=-n<ީxvϻm7ҕM]Z6)O V(!ZU18JW_wR&ya뱪õBd\ݜ'I%)ƣ"RyVV@bQdZ8=}fvQ$h]Tsff<3U Rqr\3Jd0y4N0 Tpd\90u1fḴl#<{iG 9S2Re"KC%0U]ᶍvDEaFz4J$:>\E[M|GҺcβB@z %-K:eIi$d#2oRBb74VBZK D ol 3dH*&.&jXlѲ;VȐo)HRڳRw Y*/Mlt #WK.'\0521|rh٤oi(}.24HBbNji.T^rs]N$LOntv.[2vʃ ]RjŨfЯ61Z7A3%4 \\" uR<2`1 `1Q;u_c-Zc|[|k[Z;vr 0?|@dڜROca->W+BzDͥcd/B8^Ot@B倳w+W9z;ɝeԺ[L{ҞR=~JESW_M\^UB_.8G=zQU*[/zճc3Pt#]h1ҵL뙘qJW7#~rq˻Q2|n`YJ:@eJ++#rI³gYniiIiq䤽Lts!̒QfqZiDb%mxNɥdo=18- &hu#؟Ҟ|7{!Ml[B~E_GUNS6G Lz#} e62bN1-j@LcNq$4[3@*V@!l[BWL =-Q^kfB 1q/iZ'Z%7oq՘t=j<֯ vxE,R;l?BH'uV?`:lR^JFQ!6"{4mG)b̌Lk:(]NYlc)r-tSY%Y'I#uR49# @  e1qpo4iƩX(BcٳBo 3XgF97h{ *7M Gl#b 2aH`l"R)lMș. 2\ pPq9iRiRiʝ=!/)TO֯l]혱MhU)MÈݘ8Oql j7=j xeE0\1gQ214`!3VR]@ Rip,xKB+@H&&OEđ0>p>fY :i9QT-ؘ~0"b=" c^i8/%Dh #d#%e >XGq(fAD<I%Z$Yp0 #bcF:1󄋳U.6u6%je=.xg !MF^ Z#RJ/rF<9Hz'%sb[1h  v "iyG< k'[QNG8ޏh?\tfUtj2oG23U nM@lxGsŝ_g{ܖ K^ԻP]O'_or&pAa\R1L *;$iD 8}t.T?.0-0̥.=Vv7KW ec5R` ,srLhR"f=8+׌S$2mp&6#Iչt~g#,-}7aɄRָ dRh 0[WVтU襫|P}dl9S8'^poGO1bp|:l^49U<$(,J!yZ &h-2BVѐI4(g)fx* 41Mӥ}V\(5}Ot26>ޥ:x}ft܅9ϖ{EMT琯]!Bk6w$ pKUmҺZ=Mxvu5fR[v~ JZׇd 6c>F <xy<)\weFZ5W_5=w|O|ϟűzls my4$Ƚ] 6cFzyɓѩJXpJx]e*H?-7ߣӌtf2M&%]*g$5m= :2NUp֪DAzlvF0%32RHת\謏P(Gc%'M4%N; \n~bTzEsݾb+N,xI/{Ti=Mmvl(HLWW,mUgz^ʬr#4p4M_aƈ wW,^汏xRI'q#k1,dkݧh,H&Js(a+0`}GD(ޤށW3#NYh/;r0 ʾ1wjÛ`O}h`5YX&b)Gs$&X@ }›LYiɟo7bHwLo$m,:X^_`VWf:-n074ArP7 HZ7`I'?zCpE֫U+pU孇"5•-ՠCpUaW$*nAKHWxw Rq[pHqգ>\=NJ+|\aW>zc\tȺWEZWEJ=\B_2"1"U]+E3l;\)p Jo;W$0leh;\qygU-m"P=\BB.+XK \q \ij;\){z=p% ;WE`U-ngH{pUի+lӖovKUj/[nKbL^Hq?7v0M#;^1;v₋wiaz-:TG@YG?v~nL ~w z̗dN[`%zOlN⯇.S\H3C4 V(?b\v\y2$r% `@cH>`yx}M-Ve钽\/tT~pTea5>Z}'ܪ y:5F-!x_exʒ]crtiB,^UB_.fzֺGԧc.+K۬+YVCmZJf6z2T*[2"-U/^'\ٳ;4 WVwkZ.^0\30fF:9Tfp4*#&$:&S9]AggV~<~ 7M3}lo_Іu t}IcQ5O}V)VE2M4L?79va8}0,z}sᄏ2MYл3TI$l^ldb4)9=Ty 'eީ-#H`d#w0-;Va;4e&DgOxԩ&k6F|UOϢmrM6 I܃`^(JfeN ]"(/w^f+vY45r}]|7MҕSZXg㘾Exvd$7j歓GmBNJ!DiT|qt;``z 4E&KÚݜ|OZ#vR=cGZG7H}m fQUކ6t&$$XK!Q,KBJcuߺq)hK -F[b>iyZm$Bɨ|W!'c#Z͠9[0f %xDm ]Vz$ZGviE餍U(P!6:KP[r=JNwh A^AkaFh. h;ovlX4 6gPA!V/qW"Hmbu2Bsk(|-b"ΒuQ6`Ņ%̆4<ƺ2?K+5Rl{ XUS  HJƬG6lmCc6P0w˽AAQU6MsA)%x X5lGD&0VN ddRd˵LHpm+xNiI +QY[4 U TW7"j,8H&[j! ]V۞dWNc@!7CAdܡQ l[󭋠@zIK<7b΂G ݄b :TjMPI!0Ej>dT j;XPfdBz.Vz 訽Ŋ2YRPHq Ls{N0 !-ڳ;!JPA/u: qJi0Օ/H!j{QQRA}k f1J%)`:ABPj]JDA e͒`3d5>! Db$ۃB6rZQ8b!z_4yA}paP/yrFw R5xlPBr|\1f>4Rt8!a%G<]gTOoAb^?V11󃳣u`mFL`-$ >:gAuP<l:@G6^&*fFVW 2r i(yG a;XP|Ѣ`QBą򠷒!HDNkk*fʇ`)C#`8ӥ`)`^(^b u6B[!q2p >U]XTBNu~d}:MOU;:@vTVB]+!Hb!ҷ5 SӍ_#|vV8Y:Z:5WZsLڌQg5rFCo bLfp K,`r5wn+2TJ`]PLMYځ DGCPCzuz\ ;fc_XA|YbDuN M~k!G!/CE7 F-)Ca E5FHKCH4p۳^:H%-й ᧀFSaМ6nmVܢhR;5֢Y u(P5jI:Lj='~x}AoYo5kPqab2|ƛk΁ >bѫxdPi*`)o"a2نY3 _B\ RM F:,`GIzӠlCi#8 8%oC)F @o:"@.* .Fr\U˥11DL0r";@jҙR+ބClN](Ijl$dipM(]*?#t=AyG-F(GT1volo-L2v*äIfZ)Ru޾}J:&|Om5 >6J?pޙ֧;ȿ<ȟ'Ҏ' SXis w=E"^ !()'> H|@$> H|@$> H|@$> H|@$> H|@$> H|@]iEǗ_y@@e!P?z#> H|@$> H|@$> H|@$> H|@$> H|@$> H|@С<|@Lf=> uf5> f}@2}@AD$> H|@$> H|@$> H|@$> H|@$> H|@$> (fM> +:pSX$RG(j$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@pc}p3jʭ7[o]{oW]GThE%ljlK3ܸ-1SLlKbү>YcZ]0l:x]1ZONW@- ]i=rkn}a nx!wC_pP&_tv+'tԡJ++jbQ;]1J+CZ5n5tZF]1ZQ-tuteɄhWDWzAkZZwbd"bi=t7ףbpF9ckE~0wtw݆+{L6-;<޼7O18'9v؄I)Vn_M;&fƯ Nӌ!Ҵ.눛әi$G3ը+Fk~yF)K·HW`״h 7mWFIOCWC_iҿw7/]r7o6RW;3uw+/tԡ?vi5tp_Z]6}+Fҕ DW ؅n-th;]1hCvj5tpZ ]1+(pXÍv-thw"RBWHW׌:hˋv=0>zws ~kʻ~}M-#Mn$ ~Ewler;G(NhO7ZJ`jj3Zj3cB{(Cͼ&[WCW Wǵ{ Q,%"]++l׳pi5+ߌ}+FU~M7ؙ+kWђwQ캶A1?`Rtpi56]՟#>Vyk^v0]7 /6ڵʴg]&"`2v5tpɯWNW3JC+jEtŀ_ ]1\Ok+F̾ՌRFmiyNi5t+j5 }R銢{pZnzMQW@kwQ ]"]9O>7i3x1~dWǸ·CJʴZ#]'süs:Y-;/wwG#5?/ -_4.uaͥ_>i|텉cvo_~ ijWd{ߞg@&<ؤܜ%įsuʑ4)YݦU']\|s_98]nNO?{?; J_/M{+=o3|s5/զ O>9OoK#B' kOc:6gA{/>^LoTg|5XoSK쪎]]oܒ&\ʚrk;Me=\IUZy~j¤ zC6W5R޵6v$׿B|~TA vccNۃAKb"iF1濧P_+j|a$6j,0#:;4' !rcJ2 g)LQY+7E@F^i& ܮ]ӕ $ˏf X!>a+Q*<~ٗ׷_?|`ׅ}^4z"5 ncT̴t%V1C"ОVsħ#; ĢȬ&#f'RĬ"7ګ@gU+M1;;%#  hmXfG `L"2ɼd1JQ: Q9Mpӱұi 4=U槾j][XPp'qmMyǿ,HRLv|gY#,vn2B?˿wX1^nDZ8gp5-]DIb)h: %+}kmOG^޴CTgl#GnS_JH{K_v'6׾׺i\{@w/VkȊ=QC0ٸ(u>6/ArЖ|oL5>YlV%#ؾy$9LWOit6f?>6:mqATmh5NϨgh-$p[I6ӪԾ($ a5c[84Tfo.s!v5Hwq쌹W26x9D/8 cI/aIl+yiMSR Qx-še`_G:l6ʼ9Z<Xa`)H"*e:wb"Ι+2Zpk%F渕m#Kmr0i?K$qQK~7F y8?^hQZEyzqYb"zs*WZb1S|6 r}݇mSFGqۙc@[of(q/$-Z'ۺ-=X|2],j;۾&qca#zקwiysKۃ-;^Xuw Qסz3v2_5okxs NhrCql@;xZjɎ#,8gao?_6#}lsmeL|دyn`]M-;:5ҟuSf_9,De1 +lRł8<Vi",\st~; =|mjDe%YEV .+TNJD$S; !- BuSy ]0,r`Gq>h+}t(ԾD{Lg&3ID+8F=?#'ާ[tPQ82QO"E-@*X ėjxQeUЃf6+s0j!'F1`"3RH.è 9Ɵ!t?*nSLF8b3M07`m,9-d0}|0t5-Yc2ߩtlyR-~?۟E\v2w ٻ~r`͛_Q%wF{Xv~ҿ\_烀oM'/۲?K->Ol`<F*iNEx=˰MIa4}e3>c}{)3x1T8i+d %حO(x"hnQkMoWRlfY?LU^dì,M'/Yt֋& al׍Y0 O'$=椤G/SsȹI&,={ڴ[w:?o7v?;› ›.fsOYo ]4K6i J4Lq@˖+TPiIʹ@  tls˟@ay*“Ǫͷ]iR9؟ԝ~Sn{'|DC<"22{O'Y ٵNoxFq2КvT[Ƚs]r< N]6fxro֖{ w.z%JEFT@H/ӼquˬRaM1 Lǣ[o>a~u,cYf8"vz@o􍑮ŶcZĴ$(n3˫!=>[j0.Vy,3- -rk@9X%iKlDž+cY\r Ve8ۜW !%YxTYCʢF%-/*_?bR”2-w)E\{L2G٣HO/tj90N>PoV 8S*$`)1&3XZc jaudp9& .$/ń"!Xګa6Ծ'Vs+j:-M %S &'o=b4޶[]rO$6 F[ԆV%**̞1]. ㉣p%cL&h%(xIRB RY9 & &IfZ!(ꞋVfa(7{B Q`S _#ٺd1Rfʈ]9ۏGq<Ԯ&jw MaEsb΂b.B[4dXmKw Z*_6p,7nlȊ2šX" 4G(}*@Fu2Zsakԯ`<D&ED!b;/1n^hk<$nE3 TbVbDNrVmH6\F" I'͍L6C(b*#bG6z*):IsqQWEb$-!dCd1PAyi5Z)BNʈ<1Q&ca58<#@؞>,E&EEx@\Jя-q57"񨙸kY!쑨L8Gcd%mq3ׅ='ñΝޢgBJwC{!]Iv02^zޜ qc]C^BI&2Z˘B.V:0Oˠwm!Ǜ䍋K}d8l" Z& "5xz6̋;qI#uQF!alȊG/駧CgƼZ0̑{ѹm=Dxd,nd:єů)&K+;j4th(ݓiZeX{[ >c*gv[D 5 KZZ͡qζ4ACdxq'^eYE*a4PY $;E K(_Qӥ|V(59A:`_/7D!x>*s]tN. BA;5C2 WyV2xkWp6Wr-\_,F}nj u˙NH_=_ v!ԛjGi! U4c˫&͟"&K+SI*KM|O'ґ#knK|O{SG z>ۍ5VR9" &Z^h%̻e*)K h„sm)|05u)Qȅw=eR11i5k>jKsKp+hɦ6J{HNBKJpIP973]=YYԒk5,:OnUI|gQNǔ%gdWý*"2 )IdxȻb(1BήHٲ:_)VNUbZKa1 e*+@2R 3"9Iq3@ _/BI7lrrv֯*w=Y7Ja :+FmJ *'.ĜfEuA9( j&a;O0,Mϵ4m2Q2J%+-qLP L37Dx"m(P@d0 t"LL)Ig\1)èQ*'Int#r8v$m>K_ ?ݰ4r%#Tf_řint=F[Zާ{rL$ǻpd^qٟۦŝa޷1ljx7%MX;-M3}$e2QB9aVSt٧a:_0rV5Bw svI6n\I?A:~(>ݲud#7?j Ջitj"ůj&::ZY*4!Z:"߿j Ʊ2˪]߯a4uy#э:͟GT?o|}}a~US8 ̹dN-mH`c=K㗂u%5yIB]bb][LsxRs "٧ECOWe\S9FRE%ו&^u#XeO_O-/!Mf[<ՄQcNƳۣpr~W߼yo)3ǯ훯aU |-h?= %pz}yWECy-7)tY5ޞ>VVhOrm'F_8~s2 ""m끛-Uspč(.^0kmw:5} DXZy)Ā .k܅<(:MiD{{RГ LF U$Z*sR>I F9Kg7InpXP^UϞhb{D j@<:$(kNc4&$3u:יdKjLl1#oo|NϽvkv>4z!V ;q1BYdAotʋ;V/wʓD`xA9sidQs[n>[E$U]`jYNBwk<#>KIKYDHD7kPP-2N$Q' 1 242օaڨ5yEHKc.ORYvuu<,&'^N:kN?V|ޛ|k?;x$ϟzNϗLrc$>dAJi"-K4Q.GaA2sm8dΥ>]k^a|@?{#JQFʥ~3Xpy<-QLS ˈ=>"عoB=˃7}{)N|rqyr?L`49Lk3~(E7H ELȰOgwR{>sp4SSp{p}`Ֆw ?zSZ{fjVl٥헵ir4ΏuEZ;Wח|ѫ͖뙏PϷV1BG0G&חKH\Nσ۾ax}T>>@3UϮcuy6 RT Bz*u\`7Dk> ɚD q9 d1`28 ^{sQ,%ގtĐРh<8que+xl3Zf+w}aLC}Nu{Pgg24\F$GOI-U>4.xBr#NgpTfo*YJ:(ϙd7'A*`&'Gf';T9 3N9 /Z+v~ 1wN1 "^ a;2Y+)\|7"D.MřaD%*%uցniUQh;.*Ue $峧n ;Lҏ`s1dGN d.Ơ)J^8)DJ=_nIq/9n[~XN))$WߞDM={.R+-Uā-)ƉkKqމFi'=t0ܒ k^]p9Cb]!]YPΟawGWXb *Y ]!ڧ?~<]Jp2v9$V0MwNWsk\.vKWJjR~ѕiAWfmZť) ¥B]!]1fB jv=+nms?]`HwBBWG Q6Tց$$]!`I+T)thu Q6>ҕPlW g!.ջMPF~f2JĈ#ɏ?_~~rr5{4BkAߑe#jme0Woo_o)mlçG?s[&,! ]2q 2&|~V~珘R$45X@'x2̣̒ATWgQSl|= ?eEe^߮Lc' PcsvPU@bᙯFӓ 5?Fs9O}d,w~0180f\p2JidGuhXܢ;[NerTs_ gUe]p9&~\ڡgg-NQK2K gQc2䣖ϲZ>|LcƱYI \:BT x` "upQ%Dž BIeTw~'WF.w?]8iLXMrJaƦzX`SϿz,.&2;wf J*+,#fѫzUI=P/ɥ?ÁWR~Q,fM De+K,-t(@WHW(Ul\MCÈ {HWR]`jy1te \K mnxV]}1te7z{H(=v1]kv0UV(te[ЕjۮZ* +,T1tp*=]Qa]!]1!fBܔBWVؾf=+N%/0U זBWj=+A5UVBBW5KW聮$5Ȃ sS ]!\Qth;]!& 53k&bJ4<E,-@s-.URAH[}VŤ`1.Amo(2'aU"W"[ymd>u23 Z.FcrF6D>JQIXAt,gCjZ̆>բt( }+<,gCb+DkHFiCR b.#Sw73]U|V(/٬9$L?ot;%` pw+]DvW*3A6ٳ]maK+,/ŶD+t QJ5 k&+k,%}+@ijsF5+03(K^ ]ZA{ "J+C+E +l8+%Bҕvt%a%ֺB;}+@Lm>ѕЦ؈#CfTZ0JzKE6&jWJhW׷kO;GTq݈m:#W:5fNa&\R`mU15;r%|#`/u3 %m`Jm1tpY1^۾F>ҕ$7nf;]\!^Bf}+K0/kbܠFTt/]OR4Q䓺boULIZ _%\I` yl 0L2;N g]o[e[%cu=O)#YAmQcV[+Pc>P\v}VrELh ]%7ZNW %m"])m]%9W7&h9}9$m)ҕAt?B:zUtJJhu mۡ+ԓX[WN< 8r0HP9"-]wX50͡[}Ce*-]=Ey7UB{7Rt &U,Dc*)tЪ[W e'գ 5LpqcbW -au ]qfcKx*܌r@!7~1zVW=s3M|A76ȷ\x4buuteQ,-rڤu `\,Uʍẗ́d5`*CW .MUB)TKWOIX5ZReJIEkpaUsW'? mS+AthN^1UBk(@-]}3tEzAloz,: ѳ6wX,k@WN=V: 䍡WѦUBI@[ztETm]%p@UtJ*UBTKWOҸ%5  ]Ey .YWpQ*&NWpQS+4xY'ip.%0SUL[7n-t (E6 ( zA+WO X%9', F|ck:=i3 dK!s1mMVM۷owKkY*Ee0XFVs1D_~g,k`a"6=YFeRiYZ7j {3~Jge+7rASy­k ~1XsZxkj:!b0usKׯlx3]O1]cb&kA+. uMnݓAFOЎK㽓"JNaܳh1\@r?[Waa.qPS0:,>$?g 5Ab V2Y+ "RSFDD b FрG!eLD*N4/|.pd6ģ`5.E/iA(fB^A+JkղF.F#X //,|~~)Aa5)atᕙ }d>&#uݵ+Ojop6èu1]-=3yFT}Hf͑n^`B0 z/<;zXt4 7~[L-D2TN,|?}W/]W|_2hP?R"1W6-Jq`q|8ezqF^ȋvdjG^<-X+לJbk-Tt΁1G0Wj"a =MhS9jve 9< ^mXٴ6WWnD:>xc#_6nn?0DZwTn]uWw';Jfκ}Aߴ@(0PR'؅ yP;tbh|CٹlC+yGDJNcRɜRafXTqXFI$w{w>.BddeꥬʺS.QP* yy`( d#!YȵOO3ۥ鱖} +5i\I0 UB9@)avi*RŊTa:JC0'`\L2 "MBcZ )VL E)V5w4ϗPgۃ]z!TuAvmi@C9Ae&3|b fQ H04dkT()lb\n=̲ 51"iԮãfQa'>FR`@ t&N J1u3\tfs a' t`HVtªP& -,Iå|7Y_2,IQ͋,,Nף/dD%go &vO "PU \տk TH$["`뢍=ఢОOu =B8o |\`>e#VXOXK6qp7y,ꍾ'*@WbApU{saH|MD $J5ịUhZl=PFG j@Cb x:$ד~,o+ZX%IIḻ RM9nPcaJca౰1c/Y,)Q*{2掁&eX[цh؋.wn[,؟Ũ`6ts4 lVk >~p[Ơ0ZJ3ˢ3X t^q KA 쭌IB65|ŢT_`Q q2_=w؉8mo =r<6LscUˈ 1F;aeIbCK G Rέb40b#9ep4j1mFY%e#m~#- 0|wT;,v /ģ"%kqeQ e,[,&H|F* yx:zvRO<h`xB&0C+R#FQʔa`QĽƙQyz) g-m##3G ENr#, *%=*-"H4"G-QP}M52'ިNkxlP֫pH]2nP0 + B7lއ7NQD'Qp %$e"%7:E8ܱ#Xw y\I1n(u+ӕWg LR+sT?ЗѼ?f/I8o.nً0t ([n?X/7##Dh6Ay**Y(+5SE<=Ldm=eu}ݡuuiGӗd<i7Y5gN'~i:̡O)r#]#Mci1FvX}YYe-&qr#90XzJ¸FIvg:2\EmϳW0?I5ާ0[ 2]u.FS,ק_rfRTҏƥֺ.<@rNs%gS4a7 >w=3g]7uPy⌳b_Hw äG 䵲$N\[&Q#鑵.J=gW?LzC[z*:=H>I_p{eDY-ǯmG?<&n?#+eHX8Y?i i8!tQwj>ϧEZinG׏ҵ֗ ˃xbj{jzǝStGۜn 5hxbhkhGpZ=Wn;:6qŻOmE-_4#,sgPaUFZPQJҁ e] %ؤJS +a MR"(XmP=kqm~%{b sg2I,n[U=8U664J]r4Nyɍu JFrїWyyH5L'կsъHPIg7YiZ@aL1ٛH $cӊJjI<0-qG'~ ڨFIS p}eRSM=>WR;aPpN6at%2Ҿc+&\-b#-fbssљq%OƎhiIڵ1+5|6W񓾪̰G%Ò_(޵SzvBd;Z!Z ,!:eF8˓6BLs qhd)s4>L(SVI>v;-4C[4̨]3ОSk@y?<4=賏M"^:4m)~;㞗wKwkӰ~˖ϦCmbT숾%1:q+jJYSe[@o0hst]6˰sfmںLJK U5_I,c3kguc]Mrr Ѹ8ݰ4wO1 ,K++֔hU *YN!zӷ,1z!L2ac2ac8Gsx R`%"W6:(!0+昔\JZe =rRyo9@"\]JGO/jg&Ξ[!K:/^bfs yLf`N-gʣI2PȌ9%D`t2 &E : +՞kZن2J][LNxZ$FH\{1 C Z^jN-MRӵyף]Ԗ\.z޿GGF0Z =cmGJƘ0Q 4R+xl%c) &9h2}DiĘ%=)`@eb^D" (Z:L UGj\Vb“^_͞f.f;onO?y /yW(cg#QHXؚQ0i\v @ #n-b2=͞Cؔ`'K.v m`XC,ٍ~fY :*a5qva/r` "V""TFDGm3>FdWi987 J8J ["gL 8)fmD.8I'͍LitC( f*#b5qȃ1ѫfd6KE]U=.n!CJ3HXJ"e P^v(EIOz'%SbW Vz"xkz7xý,twNp㙼O}}3+&K40\QՍ@ 8hXCRK/{`yNe%ɮxm&#Q/@ -LHʓYb=P ) ҺYVʗ6& s! F&}RK݄1 uɛ@C:fJ3fumI=4I{gJ&S3Q,GR j !5_g@ukɸ RـWC&ZĞ(ʃw<ʃw.ʃw(ʃw"8z!^&DU DҺ0߳*,yЯKae[LL{K׏r7oGO1E1#4$|Ώ|B3{ЖbaR#;/X>3s@_0ʓWWF4(>.:fn|;Yz3N<,ſZP'N &8Fmy  Bb\7{ڰ6,: GuDyD܋`"J{f(RQ9x̐~R>k!YT><{}}z u={!6OiVkn\۶w]z$ uCͼq;:]n\^OZ6r7=7BY٣絖dE2Ԏ*A2kkGghs\~`͒~{;YEWWHҗn/IS,zZ4!?_ZZ]w}ԇ\vH)D-c  ,^ \K"ŮI =\=_=颌tK{vzͮ'.煫I+X<)e p=\9..HZlHmWoS(UX"WEZ]"ZNWo$C/ H`%.]q*BUR \g/I$_v] bH pJ=\AR\( "WE\c.]"U=\Eiלփ;x渘M80qϬ1zzG>(03?~7g󦑞^gxŀ~P=73Ç[i.a;Nԍ]y7O4?>w8z/_Qژ8^x:mY{-|LSC\3Zژ >lsMht,KZH$ uJis>LPFKh%!ϱ r-n3-C+AɺN* +k7C+AIt*l]5)M[+A~$VЕ3N_z/\ > -/sJW6EW?ö*U_&' ?.UT^s}G=>9s ^ ,'ztS"Y+Kj޼}CLƼ=mm;/'˳;@9{ݝƖw?eJ)]n0)Kǰ3T}0y}##.ni5G?ěq9dst񁜺t>|bPP:fh;nY8EͿJ}b|}p|yA|??ۼϑr?GQZ!r>aC^P32uJgHņf 1~*O>O_ϘD^G{ho |_}w|o7p7md(HFhj.\P-[r>'ܽz^cvTiF. ]&򌦇Ҟ0v6-ua> yRhstsm1"X"/LSm m.qh9'!rX6}-6QO vS(X֫7֚M?zPR~D~X.Ø3$BGCzr!\fH[+bfFʳRL='@x -3sb&Д3F h{m)ʖHC:ølC1ibHƬTInpm47w&)ᑒ9(O@Ј&3Iu9u\9)mdb*ͺ --Q^R[cr2BȏEVfUь:9^Z3zosf b;FL@r3Z-E u}lNkaOnadk;ŖA\ȾT0Ǒ֑~3 d6SfTkBm`51#$ڙNZWSn\8U"?Bh]=Bl:/|ҪyQkXmCvf0L,Q vIfژRYcQdB` :`՞N $c.ډ03%"_V&6HLK`L&X@o1س;<쀶 i0 mh,+9$:\GPyUjҕA[]^ɶ 1Z](.t-c6 d ͎# 2UA$7WG$uc6 >k5ɣHCEhrGٻѡ**eX^bNqM CZc-n@N87P& $!$m^, |Eŵ'al*Z &u|3sdRL;dW&o1wB!ՠPwjG 2P(SP|(],AH8;s)#2} <ن30LՊGMWQ"t,,x$:M;&q b`nqL, sbAhznL /-H2:@=5ꍁD]`/N0 #"ڳl";#JPAO/uV qitUH1H]Qv JHoG롈lVd0̫#lS %DH -JDA EnMY\Ȉ Xh]viĈV&4FL!TMY23S>?\_]Ŋ4dyb69#94TE@ E&rZgd^0P>DTLrgxILPc $k(ARGm!@@8Ɓo} ,P -w v<&{kl4FNG/u!tC?0dji&'Xk tdDrHc.m,56X9(P"Qr@%w5uH& 8JO5Ҙ j#`U-RP2vV uPt R(5bPK`WՌ xXZgeHÙPC^B3v`mBgU酭iFrd@CuqB&"]G 7åUeXJNS NJp"*iBPˉ;qPk|¿?Vڳ!YP*VDxQڀJw[7W3xuBZfF[PP}AώODVw30^R@2؈q7x7~kp_8>;2)^h$͙#xB`f^ #w`\cV=,FZ֒u"e۽6~(N zry4|}UeW.EiI,ԐzQۊ1Z'69R|Gշn֓b  Ճ +I$XÖ:y8H 9N)Շ)6Dyǰa‰K ,j12@u#/guk[s .T:c*j hN~8hcaVJMV =jUTAY[nȚɈm!+B5 tMDܳ,Ϡӥ?j7߽Apcv ܕЫxdQi` E̫NMp 0he/fA,Nօ~BS+.8ٮZ{:uQ!,eJ]1@b pQi 6|X5fK/N?.MY b!S ٱ%Ijkx* L](h`4&6~H_eu!(︧5W0.TqzltmŭggoH ehW;KZn(N_Wbm@uaWv67W?wfy:[wĵ:?S!+onCEـ?xPRKA7>V> H}@R> H}@R> H}@R> H}@R> H}@R> k$-jg;> -e^)&H}@R> H}@R> H}@R> H}@R> H}@R> H}@R^b[> $;f|@@Q}@/RzxH}@R> H}@R> H}@R> H}@R> H}@R> H}@RF\ޒێHvm>xPƠ>Q}@R> H}@R> H}@R> H}@R> H}@R> H}@R>z7Go~+_svq3yW}ͻd)]#=n(!p݊Ѩx*#m86l;}AIZDJg}t%v ܸJPf]t+u-M[+AӕG_]G}N:/)t4 ?zl ]=e:, Otܡ+x3t%pm-ӡӕFҕB] l6CW{q+tdӕ\3Gƽy?_%Y2<'Q^=Axvq>ݽ4|zvq igo=DۭW$ʕk1혋٥xb ݆:zԻY94;ivDTw\rN.-79l|6vb) r.ߜOX5b]=zjM7܎]w-%WNyK pLnBЦ?{ɧf:^`mR{CD9?=޵6rcٿ" lI|? 3Iv`fH'! ೭DIv'K$lUBzsKr4LFAyfcpi; nJ6ֻһ=}URG o{WZrc{jjv+nߧG'!-j#D)$ %킪^VSIZDW WfMBH*lD+ mRU_tUBKY*d]P՟U/gw:3Jݰ*]Ɏz̅֬EtDTSǀ7J!:ztER0C5tbJhn:]%顎^]Q$-b3%tPR++&mUKY[*ewJީHWUYѤ|I%W Y:5f݂} } wPJ֬=*-t>&UBQGWD'g,孙!KhEUB)?]^#Ixv0;ڹVhB?ajR]ZX2"q S[CW .mT퇲iut"tE0WUBP*]BR &Jh ]%ZNW ]BbR3M[DW]%qZNWh(Xs)L3A([p쳝Ԏ PrfD%n[À3J60,VW,=eՀƜJC}h6I|Xf\aZeYB E-+LP{\JhO ~( SGW/BWRQAۤ`ZCW mV6~;F_#]).yFij*ލOY U3Ό/Ǿ寂skI)8@%Ҡs=et)/QY:>@]x?a8vvǓ7 +8 l^-^IǸ(}8l-I8aQ< Ծ1E n`CF7B(yg6nZs Hr컒 ؗ}(ξ4S9i y ף@]wZhQ9T꼕9DžNGVy/.l{wwX`(Ler?.#?>K?˔R/koީ|T|3˿MI/|=͠=\w \o`rWfigHIK?r7w):;F2}OzciKaoFtejb9EFvUۜXHn}~hc!$>]MqQ5.mre[csY4a#A2 6hZ{`\ao ϥi |?н巡3 V}Y7> ՟\% MjY[!\f:6)[*|V?r0bʼe 1r<:Y]t E"YCG=x|JqnUr1!`TʂonAޅoR]M%%f٠Iޝ-mF!oJ!C0~Do8 Ij4ϓO>Z8;,(۔CR'q{@!fryqj-J |Gm7 %cLk*qamq(Қv\\@6Mb5fĉ7';` 3> 韜 k@FAeFքQi " UU@26U42(e AmR+_|;t$LEt!"'U͌]mGC_q1ڵYGڝmF7+}t&#+#uH8EBˌ#(,# iMs!2 hG9AQNp =rTiL|X|]>$bl #f2"YLj#nxBX ]=~c+@` hL@XT6`ം"qӶ&B2O#㄁(!¸JEK H[C Ҳ‡YGSPƇ@aiM'pȭq(H!zy%z)l["g*la9F{ .XDZ^5G?n]6'Z&OXEt妔x17];L1/ӭ)$X`T9# be}0D0Й<)c"{)\4PCM[GT̖V}:G:uرZʄ2EA}0A*wm\:vF:x5ܜnA컃'Ξ&A۹0qa4Fgt4L;r 7KnӢ J[XcI"W[ֹ1GmQ >yֵ`Gmñ.g/QBiS$InL?* 5)DFc-Kcó8DŽ6ȹQjE rbrX+/B"IbuٮRP$F`m6feLN.)&PWchkS`[wܘwY;s"CUK?\  ua) %)*妰 wR/TǍ[P7b:rL(,E*OcAaͣA:v/DS@`b$&U6Y/&E.r՞ YlOG'1O%1Ͻ=䉣Gv8tKlUIvd~}2NOul3{47x9nxS\AAޏeoZSm4=y쓼d|1yyWaSpE _޼~FD*] SJ3攡(4UyXMOY"Ll닉&P> ѸW3o:$:'Pw!mY>ו;n UNz䯀trŐux?GAmV)ɾ=1*Fwâ⤥A.l$~AN>;-/M.!h]_kkxMՂvLe 7B%L$ks)_X]L^y'xsE8-.<>2ͥ' xrpd7ޞlR<ƣцwק1oGl~j,- b핟yRQc!WG/@ܳ3ōӴrdF4Lp2)a|aAaAڸYg֠YgֈYgDRC3 [4ioHDjz!GmwE`Fy*y: sQl=D-b{gq@0@>,h{}]pvX!Qrhz}3~,{ۛŪOTD<%e*y?N!Sf,`"ӫ'ҫYyCKxz^EncB#:p[z.g$dt{ޕߢ9 ,fSkܙT<ZV`9/Is΀ ZL-\XJ (WKMhS&J;fG)µ~zqءfYfl+'%$}W@C)k9w.^ Cn$ F/TR:E;iQ #}"ؕJL "v!),3yr5԰S\|"\Wz/POvpX?1Ķg1P>/\*)sA)stiz F@$pEʺӽ9-kw?yrQ1kgsXեIQhH$[YjMc.OR,tviȹ&NIEᧆ{O-7> +ۼTkWwB_i| ݉[#h`1ZDZ&!Z#5 h\"2qm8"5z'%xoSnfMnou`WWPʹm<n>qmD-W]9Lp+e\:z :HNaۗlvزa"!I=@pp@qxC(%` 6 ]>RݲtvYz_cUgOOO5W_zmDiV+Xx:%~ j;P@YFqyo@3_t'Łc3r,KQ9q6yQ3 ՚>QܡP@\97KF8WJMI{_|uE+´٧Wej+:^z' qb'Dtwvq:ݨ"S2E8R{)<4 U6z {қOYmx: /d P3zpv{$n+6oB;sQ -b sa+~_ڰQޫ4[V;x8Y2/~|]̍l^yh~M>gm r(9w`|[r~K^X [8)BP:#h)4A$'A8)dWD|E] #1K+$CbѸ UI*(-xάu1vlIRE/U>ӖkT 9!NOP z˻ٶEN™> |O} B;6Li#\JK$(L9%UDJuY[zt?#$.s-Lt:HJtA K`t(.(4j!ɫK1)b6WFz.؆t`dwJOp`́ZŻGg)po2"h%TInA  u&ȏEUӫW]D](F2(ipʺ[\ZhEx;jXę@A4aȔwIi){,Nk$PW),TQےDZS=ڠS>g{:ΌU 90tD8VP^LR~7Wº -dv[p`&b<_@dKx %ntx>r5T \-GQc)h`D7R6!sQ&%4[lmV9H{5[]i)47AT_B>/ Ax9g}F`pƆ[޽~٫~W^NaFNQG`p`,íy_9*Y98a py%N~Q?J?z'Ӥ_d7>VOOE7[S'q?D}1- f[bGȍF3/󴧆,)BEFq8cY=Q҇ilՑ3f(ݣdi9x~>봨MyrR],熸4=m%D4iƎ߸!BNv+PH/V tK(1IL8Y 2ɤ?qg9dCПoKMy7Dk<+cPD]KtUi=Fh4=W#TYl/nܳNӔ]N~||a5лT02)r"uaן]%ʿeyؔm'ј_5vF!Ii2oԧ]pesOQjNbw3Iur6O'LS~kfIPT럎͆tZlLR8 T*0l92g3;EzN!KALDQ.jtۨ7t莣?y)U@gCZh2 EQZ8\XVsqX(PО?{Xnd06HW Y  ` >mO˒[{;~W-eʺ ~*KS,VY?UӎE+TiˮB;ngh\]?: d Z&l6OӅJ VhGg*WF`ȗ4n8(A'o<ya籪Õ+=2??}Sڦ$">yg,9;q+N\y*~*ϳRnUAϨn*>zDs Ҝ \s9*"vp[_\ў \sγb%@WU3+2XW\)Zt؋Wjk E&w՚{6X}UR!\)%̊+he`5+Pb(7r^Zh=>lS`2B"!={Ta8 T*_o a]ێvnQp:FG }=QY`q0ɢ\g E=lCvi[mv9BFO ?/ܫርJzyF,,LJS\XFte+Y;d@s/ u6pU̕\X:CV\Yzz?pe6`\̅b<*V%\ia=/cm!EMeXҋ? 4W|8Au~s7l+>_C(qH$i!CJ%OxZED78߾B9=>,&ݦ*= Vc(Ey_̕v 8p*!210%ܐQzĤ}`1ry`@a9Z1x.G:L}>6t.Ӭ-5|{-f!|#Սu17h1p)/w Xw}ˊ:RG6ޗ|V) Գ{0*|'+ЮU~O)攒1W:Kuqʻac.KLl:Pa$!jtKDg{JbJ,:VqCl_d<Ы$n9IY2BBTY8ԨHLܔ 'ǔ\JZe)E/AL3\DwV8b=וW3gϒj[!KJPO}9Z``R%o#CsJh%, DD ainP TXάidF O19 $i D LqT 'j5svOtYv/ 4)Y+iۯ]ɥQm{.lv| JIV<4&gO~&wl[j|LC{C֦_?UQy6"{l6qQ@3%6- SV&ck)r3}icb zR0(z]&gBi(2TmdfndUaaq(X(z,EDYe="n1IR/i8'fDh# I#%m*"Ze'm`F%g|ȤɄ@JKArV2"V3g7"%'\.\tYm2.{\6❑$DCy'-K"Em h0CN yrNH]j+x#@؎MrTxg}F=G?(* G~\>4\i][T{$*2%M#c!m4Dx XNoW-:Tjx3.4=r̳/;%"j- 8LY1L un5wx Ld:0\fهt6M>La];RXb,=ӳ2t+kR: `N.Dk-bs3+ydOeL1F3ACɀ[l#9Lq'8*?MBjRK)lfa llD&)Xl,*cSrT/lƩXJL|s;?gAGmo3;a))(ZsF`'/q1|jK:_X)1sSx,ͽT퓜v4Wtmϊ`3:~fsu{aIٙ2 ~C_k%]Bjb|{MNYmr{7\L'a@vrZ]ˉ#Ng 644S#㤔N.st7kvȊ6g^~^}r>1o*'R(]֊O"/_/g\sk>b7gtUrAmaw&;rK+;Wv&"CK4CB`"մ$3d| :8kuDfmS29!!IVLEg}N*|dyZ!;.=0vIV~W_NVcъ9 D s{m/gy4UlfJbzz"J G LcvVʴqC/iH]-FdƈF:'>RթOM9nR Tx-n H\3'L ʼn{s$/q`ѱt /4:DQr/1E2<=*{"AaQXVaBwOge(l&)QͷUb{`~uSQLŁCxɷOLYI7ɉ  #lZ~6xm?%AԲ.;&x7" {׍?ZztΏB׃浹۔^/i"#`Rk"r-hB,C]"Ib?yL{eEoڰ 5V/Ѓk9XZB'moNo4~ WF,fC/M 'J(ٔӐŽ15cPfλ^rNud,#S!ĐHNpXf1ocI)uϝ-ͩr_N?/mgEgCBlj~ ;|M7DsLZ8>%HN\Rb9rg- cH{,wV' uIp8}O r-MX=Yo`êykg;ݱm$󃓛uE,O,ѲY?}0eM0@i#U)秆5UsU;!n%T tB[vϩ%te 1$J%8IBXiK9iBa-߸yx9-VOljW`4N0&m\1 wf1$DmD$&miQU6YnZNA,X̫zs1q}?oNMǎNw\wUWZ,`ODQg3VҀ6|/թBDύ 5DXc 6[*Oű*@tT>3qؘ8HXzTαO4ђUwUVB ɒxX P U9)>e!`- qwO 9阸H '9Z V3gOayr\tKI5c{4ڴxyOĖtMLǗz-:M3c&f5chILǰLf2CHIҩT2Y:GK1}X?܋Gr!Y1pԉ#*DQh@:*XH3"9I UfJ[v¸v@Ai5UwحɪN=YBi#1k6VH_ EǜfкìbdYI(Kk-ML$0:[ -BBg MDJ؊JBpKkC VD:J8K hMb]`\V~rsyUGdY+•ڻ ʶ4{?{6l /;t`,fIp3$0ku#Kzjz![h MRէV, SN.?𥉁Hr͆Ȗ,'a.h!q&+ۏ5sz( ɹqod{d)456VHFD0s"}33;ܟL@սOa Yi B\:$U.\.Kh;.Ó$Ea8M1Gy73 Eaz9\f eU+[Fڬ.XHIBno9jQqKU&+uǓy ,Tw7~MO M0J2a|rZ-˶]1 MDfhA0֕R%1~{Ig!zm1QQ)mNf4\z,֟Q9R y%ו%לEd1 >?.z6Y8{7x ^*dS)_>p7??7?u|o~Y/R~~ F;Q/nQ4h(o^4Uly|rUU^SŒm-bQz:пO\<ŮGlYΚeLݐAg~ @ڟdzEk@˙ϝBSjʈu}E}ެϕ$kc:(GNrRTHFHh(PE#a,``-[m4: sˋM4=,ޖ{[8$TXbB_ ΰ1“'뚑u$S!ujי|G-&j+y`&wi;bSy;ϵJ'w[Ԫuy\wq!\sey]i×G@ʵQ*g er+gI1Μt۞7b[0xM

Nߗ%Hd0O'VkBTg; *iY|4 NvLڑ4?vd4s|0;s-(4q-ʐy*1tny$XLjAuZ.ZZWo/ԌIvџfa1yUX+XGWMWuViyXf`k1qV{1gz4‚SѯUѶu*bTfYx_׏![6m4 4 )< |( 4YvwBu}7OmxSnQ:#va+Jj2JWIOp6iC(KN Wr4oiQ[{TV5 fNիQh3UCUЄ>UPgSK2{/1#0X`AFBx&ZvxƏ=AzƷA&P6 fB4V:j3"&c&,aɠVF$!AH[1bR̠afݱ(ʔFXSLTcO=]ϰ!ϥ `2,H{!+A,unQ!r%1t @RZʹUZ̜V}Pp$UT;Fc!#v`UKc&UgM]4ywYxSWxT yU@o=1HLwqeQ e,u[ӊHKy p z5h<=y7F AYL`ªHqb CR "6Z9;+h' ɵ-Ǝnǫgf֏z'^;ɍP ш?2mT3A?!.s1F}#-5hjfnԡ8$e>P[)Eu~`j-@֤/AiW|'#9Ls":XBRF-R{S2;tx4ϔc+#]CN&OW6D`-s'%Ι*WH *BߎOi^I8w3Ca0=dӱ;kn}u|FXBbH2A ݠ_ Zҫ=_dU8!dy|cQw3e.zgۤ?Ӽ7,~-C_˽x:)oƽt{<2/~r#}- c}ơEFv<}vp}AY?.7(ԟd}Jr?YtTI,W=P }#Gߠಿ_y~4.90E8v3X4Zr;Af2jٻ|j_=yw^}j:;SJ'@Ճ &8kz~Sk{^o?)DMWe\Q܌PXbk٩ijbdH 갪tVZ^PҮ'r.,6G !! 2 xY#8H4?' VߟRM[v5iv?&UЖ։LbD#w0W6 rA1T a4JvW % t$oTdA=} S^xjpEt.I);Nis􈧴Dp85qy1ٛ%aG>@3E/{e"{L>"4E2_F Ͽ%lvoo__xhgwC)WH)@R2qVLW,|-ʃ{?TxBq.$EMLnգWO,<h{ T -ޗoFVW1/`QY73|{P@vmVZ*fS2xz~Ҙ]f~JRǼ%?1rcO՚5J2ƹˍb8g\Ss''sJ OL5'U򃰼TɎ BHH$T E)a ix\Lj֕CQ{X"d,!c}JAsZ` zWֵe=O^q>eCԔ cfT!L )3gtr,icLVƴ7Un4gbCh8F'Ġ,ђP$u0^)ECHq$\@Keaƥ?'d@U.!FSU8Rn+帓F%j vt&EHZC`P/ 7po bMyL En<`l ߵajⰎ,m;-F FDh(MTZ`%Q#n}բS9,]:'iYֿrP8a sc1&bk[1LŭJxℲs+~nSt cF[CW .WmVUB)uGW/`0qZDW 0QUBQ@K+h]%R\Jhj:]`Httc-ZeRvph ]ZI%gdEZXu\eW#Tn'pekV(k؟,Xx!}xC6!39Ωsw!#|V ;[pٴ?(* q&fbə! s6 >KFpx d8Ln2">h/^SeW4L0˧Aܢ^?ݨK@ǩ0E*XDΒcSNxl~UqSpU;-{ߩ]'?n'_!ER+R3Ey gw1-jCKgΓy{j0?ڳdSS.I-)#*'N1r ˅ԁq~@ Bl w[W[J \IB-+h ]%5t80M+(8HWHNhWW Jhh:]%v"JSn]`ܢ`P[ NW,;j;] #J"NW&\NۡeJ07(]ᎮRe ?fJJh5j:](eGW/;[H !BWV#tJ(WwtrRn]`;%Y &URut1(o]%6oV˦\mg6- ]%ZFNW %G]DG0|A0ڧZISuG@>ѩ7hRVr-c1Yl1Q1xwꤎQaS<`s-:2]ė㕐6e1 {JكgzF_pX}!38s vf;նȒG $GbiEɪ*VW&G0JdT20gy (tTsAEe:s(SJC JC(>vIM۹Qm9;nEUDRB<WLլYSEJ!; )S \=4"QOiWƯ.OcuELE-ig-Ջ09t|-u_F6 L46WQT)EgUiBT"sM'%=;X_E0ߏ~߿Ka޼Гӏ/DHswlR_ݨ4d)+ibRjb]h5vV{-˒cI; yd&QAP{R B!`s8&8f0-S+dL%JY^KKɍILѿF fUR@N#Xr -M"p^0.ott~&>Cm t y2eNJ@*[+Ogʃ+И:)kn4ZMq!ciprroBZcL;}v9_P>':(:wwi!$/A 'fm"A)r`Yp?g\>^̖6ы'xNĹ~/i=<FZ5Ef%-iuБ%mNkXJ`P"Q`P #hcGA 5G6H!pV()u21F2{{T<6ɭĔdSxRh2N)+ \+LQ C Cc4!@9h=kyvS-pkB];hk_/'pQkg*A3;D9xP8LHkgck}I n8xdv:Y =<7g1@"HO&'rIS%+BwqX+##Ǔh G R8o2i4gI2}/@ڬY} kcI- P6Da4_jċ/=;Ξ9Z+>+wAAb`o$UX}żF:h6{\Nr1G;^KNǚߓ1}ӯHT.{.#K%]Ɣtt[JzeyFyV{I|ԃ&O_+~|:0|s}pmڧe_룗7/KSI0"goȍmw_iOkn?p߮G/b]^RիG.fR: &50mB8[<9x<iQ=wo\?THta׮O3֗KK"o?щϛAbIvhhC@Q)r+os4z@&%17LU{-=i&]eb*$GԆj#G]de@e5:W_}J6-HLsёI. 4Pɕ,8ہ^#W40jˋkL7*~wsiﮏWٛya뱪EW*Y}QYۭ~^h[ɇM7%|9ĕL}oY{$j_]ԒV:hz~r8׋_}X_&m҆B B=zs5ɓ&_ߖ._ ӾOhaٟ_Fo~bϛۛפ.'|t#7ug]:ZwG0\z1'\toFsV+ tD3 N3WZeRQUŰ 9<\s]Q>݂\ \S(uJ]pҳgP|YOs*հuEVN7NH\GTd8L^4ѨQϬRJ7 Pv%#1=yaC`MғǷy5~ܺ>R6]JyFQy> dYS(l µL)3QVYh/ُ1Xe/$i:L򁾵7 =(UΘA [ q?n+d>dCCNXFgłwa : rۈY΄Ξqqqq+V3%fGɈe 4(RDy `k-iO5# l$t` ޾x@BekJ~}8m]#|""lpo’ZRbr&-0Q4h)O֊#B ^2x-KdH $ȝu{$wS8eJu{Q\b#"(@.}r a5)CQaxۆЛ;Q|+h~0c6ri~1 S9~Isy o{ @6B& '\qYćhiD27 ]y^@aMF t$B("*)uys|'ឪPKi&ne> .X=y\VOR^Ċ[e1멾rKpP nEFZF}`q.<\vu?ƶ{]D {03E?[~>ӹcZԻwֶ')0GK:K٣M|ɗBru&۞o䖮l#3"ܧ{\'$jF»mO&7rmݍ.kэ.kɍ:)Ay낦<]Nh0P GO0$⨍N8(O gJY18`1w IT",m=3H Gs4 eo Jܦ{-o5Ϣ;x=XIW(xHqaj0fv?\#^}"ǫ5%_3/\zrm@sDNBJɛb8*|u(:|dBu Lp)LRRaP ,%1#AyÞ SP ^S8n]ʟY7*H…b2 uD ;1'Im0HF=*XՈq?}XG&`4TH.$QaX- UB9igy)?yjH 3V% `&b1G"JυƢT'[|kJwCַS9KBEVN:]_$oI/Ғ%*V00 IbeƽK/%f _'({S75O K+}ܾG^uqw0L wC60 f fjD3pQD qwt6z _64XІ IL /gP9LT'Y~.*<^̃()scwz۹.H`~?/Mo:л`inw{b|yO7!)sE.f~F/dTr|.zG>;7VJ^AuX%0u,1Х'#_1K3|~i'ǠQF_:q:k؎ח߼x˫߼D]Njׯéqpp0$AL;?|B״k&]&G]M&&L mMւ(}0y_vF.|P.n=e,B1=+@6M-W5QP;TճB(jkuE4"]~? =(GNrRT`-"eQHcZ..hzңj&CwHR( }_acœ'뚑>@֚udКM퉭_,Q^b{.;-ڙ^*ڹ3䵳V]wu'F{K(gK?o T8-%& 4?f)y*~GKjp n%FQ4!JTGXro%wQeIL<uKI0N6Dd38)&]661rI $[VvE^6n׿gT8߉s\c^Ȼqz%IQ&8$^f\rIu0i 9#r9MBR)l?0ds}IZl#ٜw #h8hǥIa%ڳh1 GSgn ;[u sX]7]zmyZC)`D%L1yhr^" \,I49GHXtHhr1P)ꪢ,*T0B ֻlbnGEڦmmQg)k-GSӘ0O_djKsr0DLMmZzSP9ھ@Ss|@ &pU"CWZ.L( W`EF\%r59qJy vq(@qU"˃WQ]\ W~<3;_P.f6j+UŜM3Mesz{~_y,f 3Α>|ʙB3> ͤdiPݞLy֙2)nYL +LGMsLѪ`@ cOvO>.h {m0y0j:зFͽh<.2R WL9M<~xY6z iPLeYs\Z'Ky$"kY`{4ayߋ R+K`RGSJmΌ6>2DcrńWvDN)cJrNHX. Q fR`\]5KTjvT;@L!W\~0Dh;QQ\}JTg<Ivq%PUGq%+%B *ID@+㘘#u:Rl* ʏ迟f;2:ToNxA{0+-d+k=F 2+T mPJ`C_UܺJ_LY7&kI:cbI2jIe K}JwV3z_F ].UfVgz( Qh/ݠj5]\r&;StWi*Lq-w'όr*^v0`"r6ӻ*9;o&V}HPX aGŨE*U͌f YywEE>JaZ>bL3ahq!uO1h(2œGwˁ^98x8Uw|ȶV[ՐglԑҰ"jMM +7iZuԋߤ]#{[}}ؕomH??/Xr2;qǻĊts=JUZ~Mj~'hzՍ^u/ƄX,e:/NfjaNtM<'kQ$n#W!ERsX2zKyٝBPY9dg8,M^h_b{4D. 8#ʔ Rz& ޥVyOUk@S,!!ɒ^:LqvqHm߇:k0p}c^ e1;;y5ܚCk`UlE-;j˙Η})wTL`)f%Aږ-Af }BR_$hbmNoI[[2'Gȫk5L68dٰRT#$N$`F`T}NhMTՌYs.kHWF{ˋϞƔK=cBJ5,NɹR&;+$0V@ꕷ]]]>x&R%W6G|f-mh7/o{~R{0LUBɺdePv`J0ƞB+h-awY䠯 dr%۽r˝)f A1.*SFu4V L"TdқaL #_~7X;pYΣ3?8h l1%e5 bQ2h$JM V1IL0Pb}|u^oN&$㎪'\W\sn[zBܕOrsl(rAJbd\CI$9F_*nb wׁSlP3ZdC5 'p>ԇ_((A_)i8p;LNtd &k]C6fL1]>JNp|a@ Q%nB I$k%ԙ|L3]ef$Qb}hʚlkwؖbY ErI#9Q 9 !fHHn4v *Xˢ3[zD S&u @LfƲPަd-iM II2FFPIFpuPՙ+"%GIAI ^*w{u fLiu1jBZ먝.ox6=. G-rP1yU&#w EpI&$Ι)EJ8gȩ= f]d5||j05V>`5/8ft]o'$l({Uu4ȍ ez4~;jȔoG%n dlŒb0X ]muH'WnU2ge_sslnf9__;>JDr[ U AgF=Weml3&Fj=L] [<#O4i/߹ Ʀ!N=i CZCqYͬ.Gd\*&uRHN-~rDԜ"Tv) * .c8w'<ϖݍFqtVp:RR% IuP2IR$=g AJDD:}4Ob,/ĸ7 ax T&&SpZa$1`u٩ȔYv槛nR-z @ipWY+}3%$OWRn*6{-=-37Y9XT#܃eRV cƸ&OJ`aA_E#3NSozZMOoy>6I& Ċ^wrBH%"*GfH #W*C/ ɍ)2/l9>xڊ\4PV{zә$8Q,]A;$sO8ac=Nwp#)}סEn? ש A:^ˆ^\ttk}firF5*3&dd;S!kpZha?~ÃꋋI/D Xʚߑ2o8h~@u.>\= FY.F`xj,U_wTW gL/3l~͓|7>D?<~ؾۃ3Q4=߇]`0"{1/ǯpF~=nDn 7i/{j_Oq-t:Ϫ @Zko$el'oS7+~eQ~V/էDjջL@ge!rGMDzح&_w@VDB9ZiQrqU5?.a㩟NӃ8t";wv>|8 (]`JaCcvJzT|Uo =xӶ˲+ݪ~Gt`,)ˉUY;?le_VmV.NՏP9놊(m[2-ɺL޸/k-U#] "Nexy;SVtj|FEV=eKp~+lU4K{)6@k!w&?? xAM㋛ BC_MhSBǾ]g.EgqCTc\|ؤ9t\ئ{f1q'6aǢfttqIr>x/NnX0zuh/Qj'fύkëxrߩnr\T>l]K2EcRNMRMt2ѥ(&iN.2$홥Ò~EBPZ)B,w<^Em\"N@B3 ̢ HtT Ŕ=^+=0fW}Q]7R]i~[ {U=+wR KIoBZ_V.MiF;ټ]BW1˯|U|*,n-v)ʢ 6eF_˒<=\+_sZkjϳRm&%2YnMݎv2Vm^&^fZU?'BA\t:u[bIkfaIthjavwvnlcK4F߮7"[ /j}tԉ!wW߷tOyԭT\uaL7s?]{zDqt3]g;O& bJ38,nU>4R*TY/'593(}]_Jx9=_% r'x}u=דީ/nL[{|! d(ˋl>GkT^5Bߠ.,G;^`s ,k`YUut'v`Z5x$CxujBt?YMpnqoA?T`YJ:@nPZ)K~wCk;kSKR_< r.@Y2!ʬ04N+H %tJ.%%sI1lJ#bYVg9Y)yBy^SGOz}D48es7'A"yv1h Z&1hI YA6$%br<@Hih.eZ3jתeၚ.mBi^r|˛DU'6]omdzg3Q\;~9IgW]S?-1-1U_zm֪_qb]GF?.U]R('E+%bs3+yȻ>Er.gb~0ywy:}#<G{2¼ A2yTlۋT͞&IAscn.#@yz!wF5m1,55~04ܦcrW{_/g7K?xݼw'#q0\o}phC}Q:0[su^U?ὧގ3|k}o}Ysf>yXm{ vfXi{6`'1AM\si>Mc%eѩFXq94E'2ðv%7ݎ.hG{&L[D.rv@Vԓϐ1U BqQYfmST1G$rQYѡ$kQJN6+k2.iHy5xmw~8X.4_,t5hOS |F/R4|"O3ȗ-R$k-/SiVWz"b(Rm ;'*8*Ju.pUw*V;+n87 |ઘ\X{jp}>\+;+a9̹:* }.pUP)\VCBU1X**<*;d%G>wWR*olƇnM5x;ickVp:tӷL藿h>[vFj wd-Cue4o?x>(kh>7@{5t,mlv =;_&i-Z+W>ZI7FzP"~POŃ{0rx-n ha|yI uʥ~O?9a)$``&sYRY 5+ʃ <+ʃ <+g' XfO&bJML֟i39!5QVY/'cc''/` w2YNs٩'خ/il0zZLƒN>žA>i[FL1`FE+=蘸??%!e[{'Z[@i4iA?.h-wGzݻea:А(?q=o-xTw|qۡ6R-u@ycbR{ҁR2\S 2,9b.?Ⱦb->TX)rr xdEZ6塂'Bm3kAgUc]Iu9pH:_/s Nt wGd2?3qs×q CR.v ]O4X8a`YHQ0H;NuLL/I{-Bd%# #H*PKM)<(?]}fRGOzy>:,:r48es7'P&o#.ѢV DD0F7I:P TXάhdJ)0A˶(XL = QS 8f&T9r&K.4\8T\sNFˆ sΞ`'WsK.v&*YЦvUZFl? /]w =0سl,. 6;وɁ L# ʔ4tsg=}Jc&'qcgW&qjiu)tH|;:4> X-o(^|?9Mq>GH5K~#A¥naQ2$ 6n(+(R ʜYnrq*mA&TwP灢8x˝hi^z%ctDJ@E'1E#IC:P)N2+e!FuAw"_G^q%N«s Hp0׈$HQl: ]eҴY19.&ޠҡ< t:>NӍlb|8FDP\8L)jSpB/F+tދ@@f'|Ze=ĭ_LTQ >(3Օd/;{_BJ!'pnU;+Je_@_?A*+P I&5OGe.|t9̯)f ΋em*%_ifӕaU2j.ˡ,{tkSJ;A9 Ѹ$J,D֖]߲$r#iXw&5N~Pn>w1ilg::yzs0Nm6}y6e-k_47z^6y-zrzF6QѲnmk~ w~K"U>0ԫ/릛;_`뻅6j}ZԇvֶGqo+0,mI['>Oչ3X]} &td> \* :؁:earNJ$5iP޺)OIK1 T 8jch-3SAh)t4D{RJ,zH) F[ @-R€ќ͂ƶA"g:\>sck]yU`']qR&^{iv0 J7U<ALW rhXczi x`!L{)c.OԠK d@Z9MBҁJd^ aWl z 2*A;.N(9UFbPSeF tp#o,lH d#F8?jjwLZ>\y,5'}8G .$P5:\ v`"@!"}*Ї,h#_>&A!6[%Ύ^Jf΢˭#p6p7-2J a@%(a][  AHa"-bZ eڶ]?X R")9ɔ((fEwZI%pD0AP8@2qVxݕNi0Tk;8,fͬ[UB1:H]!Xx%E 4: $q HHF(˿4tִiFJDM0fi*$F0,x*higy0Dvi1RˌS!:9pI":D@hL;1ي s(UOJJ|t9l]߆^4 tD9A<%8W&"I6&#[i6OB $8X>Fq?{PdU45wJK^ PS66 ӼRzK "Pm _hֱPG?fW'159]Br0y<ӼuV5O5X~*oM^^g7Q S A켚[s{@Km, s3\L 3FGb|HMÐanby pTʨ8 .c/ Gkͣ4jóJ`"|"C#i`ĥgȯ%K1>٨:!ǠSNwUgau;s9,o_~__ŷoO1Q/W/`48I&{=1/дazh6C6z>ی|qoKW*}[K$q狡_tK^zÖY.t b>Wl`R[N'nEIwר;Q!v-E_ר7Hh Гr[$J0*ENo`G!U$EBhruFI{78"//!4:#IiLVa,K6qP+( oƷ~HB\w{Ĝ#U ]'<mh8唳< N~| q*%d?P䀲C(68F81*1:#qzi ʃsrǓ圈]>lͶ;"]7C\lz^*}7É4@T,(*`oC`4?,IIhޣJUvh5Z`4 et[@D`䉠*:dZȹ⬯S䰸dK )z(ܬWt#ia;O~&󴂘IzspOo \&b^2H_xV|iOM?{W^,Tҏ'E0 ~3b6 #=OB<,oFRJ0()ei tJ>+7r*/Wl9׫w}.mUh)e#[ǡ󦫺\sʑŴ&{ vXʆzٛRKVuޛi0;Umt6u({uVYYUI%9 .+'e;5鈪"_ǫR},($E s@`B2rKRI^S92ףR>lXJ97 فOMQG+}sо{b H V+Wb8fF Z`oe$ |a+7Cl5 vQlqw,(x1S $7d^[x%X<-Ir>GZPAuN^lgo3!{9H> wfloz7XNn#H!Frf`py>TUJYQV9j&l=eK'_wܻR>3ӷ~}]Wу/Ӿ7,R҇/x4ՍѤRVM줽/ߏކ)4~T>_pLK0YI~e$#;>{<" 3n]n$e 0>Igx>nQO~/x| Cx7O{0:"8r]oGWea/&{A` vtW=>DIQCiزivtwUUջOp5PG3*+EֺK";kH8͹N)bxЄސ[SҍTuq֛*Ҷ٣ZY{'X\[&`G5{&;S3ra2qb#iH;Ϟ3gΧO;?s%3LZ7%]rX>Sa+T|.{JZC~3xXø߻/U*2' PRJ]f2z+b4tT0tg2HL[35O9$y],#5U_Z(_GpS㖾}wUu]ES/m몶*1ۭ&yzq[oq̚5U)$|2'26<,,~v9! eC6ڳ-R溺rOVΛoVTZC4Ul6%ʹu?Dm5,U}Ӆdaf&YhأwHK Ҩb|ؘw צVEAw21Ee. Q\滵v(6(Թ\poL@uKn{RۧіʠMy.i 5Gϝn $DV[_Pgr.3`QybHHfL% ;+ui?~';őAԎatp 0<ǝI%¤, "e@i*H3:*̕ BA.(f $FɓM~TY+:{eX"s=pe*É)e$"ZΩ7lgu-S8qrΙ> i$z2QUe)U,O1QU#bhe1ӴX BHH$T E)rc 5f9FhmRN=<䐤X a$#0 )Fч%iq<]SNv%vڤ0[xzdMMY#Xid\pƃL{" a!OI;IBANr<"A Eʵ|@GF:pr%7N0%AY%HI!@_i,kt#2\* \4.}MoD#EDM7{,ռ\)zUN>Cy:oCx{w+Y\YTQjs~34K̵Q*sɭ`)$3kMrG{}]o}`x05|v4q-JJV9Lހ?pbEтt jP!fH N,Z{=9@^s%@ ct&$ g}p˽tfxD%啇͆t*Mڻ[1bgw+ZaʴKhxBG/4`b 8*u(Ҭ۔Hq⦋ ;pPSP=22KN+:.BOȔ+:af$*:oB!1qV{tHbaJ[!cw\%4?_nf/ YJȏaM\Z3?šC@9ϲoC'0$6:x?~Ddڑ;b -ƶz CC5R%$;~~'^1|y7bgٔ^USe īmŸ A=M3W]μ$a~n8.))+ʟWgձT,16śiH,?IK".?JqxT=c_&-?lhӤTr E)@$+@mp`0K=+ש%qf`JsvY!ojl>H0u¢<[ue4^Z߫Nu7 g6n^Y͏[~\\ˆ94M@5O"w&񠨌%4YE5xf/7agr~81xfqV?}{蕢{6M-Zz5$<,yWsBE*5Q麭cLh[@y[q:}Aϡ'ra%M(%&:WgO%~߭H-Ag_iPk(%yi7F |{%ꊯl5Z}z56/ƽ/3\4_V[2̋9.g+tM5 .M?4?,H`A? ,Hr|찠R` 4R oF3U!]jd :%7i793X?pl):4y=|{t;bxh&IdT+Jz%XQ(DFc-tr 8M(rnޡ Ȣ`;N)rؔ)$@ޮ;gC+tL{{eI}m)2\M!ݟdt0nبK+IQ,7 k&:'s {0| {wXD|D OC +˙3ɣdIJHZGe,=y"C=ՠBJdO(:_ПQ΄6,mxnuVpuut`| AL9#C̻i_Td<"ZRbr&-Q4h)Oъ<$'x8>x,KdH $ȝu{$wS0&rEDq:21% eZ3bд豉hj4BZ5]Agxu8(+0I:Ǹ( cF$a MpFA(\7ʚk,"tPlowٛi"h ,Q/%(7fJY8PFpRntOr93tn@#aN>Vyr>I"@aeuxlIlb20OA{!Ī| fR-ɂg{ 3K:a\gή5RBW|G4#luK*M8 t:]vw9T(7MUn`2 {m[Ք*U-얞PᖤU"2 Vғ)rpfvzA[tkBnꖔ,$f{: Ā?.&ms&m*W;LF[~\IoԽ+5 wy07<ݞ;y0Ȼz-W B_Yx Fټ~StWofc>ڦ߾m{h[mRm1sPM2y瑶r{_:ԇ 䖶l#3"ܧ}\Aʥar;Pȍy 92=-FHjPӠuAS'ԉ a D g 4t^GCJͳ8`1w:"bKRm=3H Gs2 ;u&<]>}=s'w-ϲx3XW.msTr3LnaU\0Ĝ U@~D& (ZR&e5ԫm̕C(gD\nK/0 t{Ed5'6kTKR3@ !S̬)Qpoo.ZeF3EQ"B* g}!F9% EWJb8*mgu'|/VITAnמDW֝q`:a| J@R' { F5"w/2O9-V"j1HS!4Ga06ArztVk'EX9% t2 "MBcZ >E EdGXÃK;Ā?" SN. SUH.R,-En$~,1B $*ֱ %e(Fƅ/SVVC ͓2A3n߯G(]'E]uR-_Z*RܧH+pRP0E|oSwn"y1#xw:5 % '-jkrTaTp9CuT~fQL*)n;? A}fAXF.އldJCRUEyT(B"Ť_JAM^M;{q֊ds(U&eKTΪ)-T?WO^Nn^-Ua”̄3߻y.H` ~NJWf0~\1ZZx-1fH{3Z68Yހ"06*\ydˑemmnyɶV[U)I s>K i+f)|q6B+v *VKįUeA8s-L?Û^wo/5&_^]y-:gzX_!`~Y`l^H${pP]-qMQ:$m =ËD#QrK1`K4kf*UA7o@o1n55ϷZsԶUon1󆖗<0]4ހiO ^>b$YYN:KZ}:t:>S^]vtuJw57[S=?>/~PF3p3 : :'E;2 P_*;\ƭv:$T/I\ˬ9tfFm $}IRSYfѺ *ʘ׌WռT"PzU5mV3qI-ky`sWq{V;|kN;/^z>f[^?[1g9xT{gcq -Q*EᅃMdl>hGmbvJ+\MpXoy 5:bCC{1iߘ#vS(Ӟ9 q YoٲESXX*RL]DOC—0INe:$vغ,D"{IM$Q %` !ehtiKjB3:$CEˀ++RE(I,\8GY _d^?\-7Bܽb;v+Cwh'>cn×}_\$`(tdF"8T| IsjS NQڂ!:RA" IdTB!+'NUI[p3q8qe*CEco־FHkVb,,3FQ'ӯqzQ'mQR|O7yDoT3F}|a7;,M{I]咚5HE%dЦ3 \њJt U1<0ydsINlN\NxQRLCɻZVBLAt6 B.RRd@LSa<&ai5d1M{Ӄ-o-VVoz@ NPu0I%hbHZ :AHfhRC]]TfP$t5:ѶSD~׎[T+8y%n'mDcM#9rE-3kȺ(JN=qɑD%HdN "ʻN+ a**ymK`) Gis)R(13IJzc:ώKQ2QķE0β@(ߦwynl(eeeFFAw.c㙷xFI>I" (c2f}N-L9x c&Wx)Z4Q(_VfrSR;o:F[miQ@96$ :!#^04BFV 1`q@, SaP'BR.lBp$E[9*KXkl yxɔd/#ұ# "'QCf}ƻȪXXû1| ?J2S_Q xS !!4evZ{5T|d<¿& LW]V[EHVLAN$cb̘{ĜQ^g#ݱ,{:kVEchK"1R&XZq&l9̮Frul`  އĐ.3p<@ [LSPC4C$$0GFo cN9zJs(=C AM3q5;aQ$K nhA£{9Ey9&=:GrF[pnS h;Y  H2T2I-Ʈ(0q)([P/=!Ϋ;|'Q?LwP~[Rh<غ^i8Dr(i2$U8?@1fmiM9(:uɃr$RCYݛirM,^i2@|9dׄL}d|{߽zg/L];3[w0>> Vn{q~ۑOp4 ņפe_uuxw>Nod֪ܶ~z\TR oT|8"Ν ׀ #TF7"c\\ ֦-=߷tڧ@mVH"R9Q@K̼"Je4StuKc9F" SJꧥA13KǫvhqTY|n˱Re#=k5$H0XKhTBR ibB2u">|=³x5^1IѢW&f FYC ^0Kb"SۚnGpw{H?/ɧM6 fC7 65vQ[eF;?h'dC= =$ԣ@Qn R '^ x3)@Li8'Q|BU4xGV;^@1Ɵssگ;rUogb]wbIY\^L> Ҭv.?LrqA{_kɊt^Ai'c.x=W_WWW;bl;dMOk&&d8y~d,i=7&/Yg >o㬻q_ʟ{^憷* }V)~ʃjwJ+H+uOЬ$YN\8{執gQf?B:Nz.٭kb\?j[jJj\m`mTl501DY'1ns-[>tj?7@/!?wXM ڷt.\Tճo?'ӝIĖt *Zy.ozoPO䙊o/C:]21b9Tu}Zǟ"wL <˫fJ0g;WS\E<,fLi>]-Yzܘ%ga|,ҽGM0oPDr9È"(h#-dix]$JJ?Ѹ '"gK- #QF#5QCRsE%џZ"l&Α|WS*`s"뗳/~}U$^%G),1  r4dB|T +RUVWc (1-AV2J2A64C`C{Rn=yPl7iaXJZ }ЫRbLOz|ڿ&?:˰q:k3%Tl(Q AY)ז_xRDQ[۸rRR>*PPN: RZTkL=c;6ӌB>6էj7^~]M'7=l\stNJ¾Έ"1(*ZsHslPS^)F56P}:b6`CU] ѯWc]DEgZ{㺑_iYE  l2͗>'Blɐ,:*Һ)Agf^ڳEڋ=>MMKvʅ i 0Nac뜊I3| ERq^ 6bZfj(|!ڥSBJ3lpnÍ(ĸ/8[|<yfFFP_Egym@vCF@`V(hMfeĨdF.qv6}䚋rSpMK\ ed;3#vF熩ų9[<̼( /.xS'eӗBѰX(4eN}RgKolKLqj!‹O‡EǾ>gۥh#o <bw{bsӏt]빟`=5@]VNN7O="wg/RQ92^?w h Mr8$q|ǡ'F{Q 8ZmT;Ǻ ٦ln\Yg9_d߫󅯾6E!n҆OK?1\3RZ}tA^~|5Ov4_ٷ:?ǽ~q=u}9FٱOzKXzj6[)dMtꢤ s®cIZO3ɁG&a/7 eyv\,YP 1̎ :Cqhme8bujew0tHBWS'PBW/|`jw(.tJ(RW7k ӾP ^ѕx2WtZY:]ɒG~~nmp:9nQ{?>Li.Sm.sܿ;^엣峻bE?i<0~Nm_=&Lھ"n\ ;]o>h87׶7\_ :_ buzr}i̽A ~FdW&Ə8q~1|xy2y3uc&5O5\]nuy$úLc&N'..EZ5怼`ـP@{r9"Yt=9W0=hPk< ]R`w@g.CZ]FY8KW?38x(tdNv աӕqK(gpW73{ULf~(i̠܃dN=9FB= ;p ] q (YЕut59p?j ~j԰K+}8j\rCW|[Co[ ]= ]飩^bAsynmDxͮߎ޼O &KmyP]'i}{K>4n*|+֦j۫~Ƴ7-qw{jdP*Vh a~L âmo=$k`X2쇅ΗZ9n*D̹ko/ɦn2L­R%8(UꛡDe(M VW!clB2KN! ͫ]J+:!2r3|+0ȸAæ|+ WX AHXPED&T"wY52j6÷lt,,xϺ0LX;Fc10pPJ N u&\}Hp fWy(y3XGfr~.~V!7W-,@Q"Ł.GiL.0 #:l;"J` ;lBr! ]TIxYUDI)b32Ms4]m GH T&% p`l,@u%8=a}k?6ڵgM2[  Q̈KSM;f#1 LˉF$Mz݀]&.oޟE3;_ V=V~w zPZXI| xK̓K<0.VP dI] Jkr Re1G=GN`%|`a3{=h5'HiD.0iym*}5[n 32]s4/ LC l$RrNn][ q;2p _ 8[5jP4E*֝6 ne0 v*"}f5Be6 mDq{-4nKct֑pi p,J›ڀJV$㣷jx/G$XHIX i WSG#Kp03>ڪВ|0;<_Np֛iώN%LƅՁ %`;t.y9hd\i1 l n'dJY:Z5;>kM`!e05A? oUfĞlc E!9sE6* v#bosnE,WR,wWՁ,S6fxSؖg'07y02+TO'ۍ+IF)ډᓛk~1ַ R f)O(cDt |0#p C^:H9-0gǺ:ɶT?Д n FI4O *~֘M5UfT〈X֒ c64&R"8 qi`IWKj$dipM(҆O]>rͨ8 c8q?:ONί]3YkL]Cʸu6U[sFFgk?r53uFv[*T[; /.)ܜq2aҋ܋7ޣbVTx0u@@d(Z^bB{إhZꀖ:hZꀖ:hZꀖ:hZꀖ:hZꀖ:hZꀖ:hZꀖ:hZꀖ:hZꀖ:hZꀖ:hZꀖ:hzu@e;6, o~! ,`L/ZZӤVd{{NKږhDZ]}f!&OuUu: EϚ+0ꐽ( 74> }@b> }@b> }@b> }@b> }@b> }@bD}@ {B> bWz*> U1wRRBhB> ia> }@b> }@b> }@b> }@b> }@b> h> gH`t4AVeB>)T1=> }@b> }@b> }@b> }@b> }@b> }@S=\X]{j.5RrmlnV a()Jqc[hؖt.E04ԂmKS-}R`Sʶt!m֐ұ\MQQ[OFE*rA.W=e`\Y'c˳%ֻ}6媩6x 4ժÍ+hp/p=74{V[uƽ~ &(׼m0[ênysv~6#Z˅V0EeXt ' |^|c|?oyٺWۂJ\ ,(b`fP[13aV/~f˛C%4BR 7s\+r;9BQ3Fh:Q@%@Z)(cQ.J\!pd q"W )c\MkE'B\M&BZR$*ZkM$$WGHfO.WH)7#WnϪwsa Jʍ,WI_]$Wi~*k\r,WOziӑ\pԂ\I'pT i]Rr5ARN( ɕrJ*CFDWHkCrZ8 ʕvF(OH҉G*rQ.W=ed\ ޹4\MFe]!e๫)d_wuwsV^7gdrpe񾉪~qHak**n r8_v/򮤷WKxNEy^^@ъeUmB ~zX3Zc)ڮǍ|D SFWIJ L SvMլk.IVrluUjUah)Ti[Vɞ+JG_) vj cJQN$|O6{b Br1z=\d殐Vg/WH9jRrICHɓ\H##Vb<|a~zP95(Iթ,;ekKd=ز+9|Ri](hv>f/_(76jc[ph 8YTAH )fFg4!B`OE2SI@]R)UpzTB`Br!P+"誧uS +vΪMU\!vRf%ar%z9Z=4gXyJÍ#/2OuzEiɕL+rԪq\U\!؋h]]2* ʕr9G9\3\!UT i]; ʕEH8\87+f )lȕ\Rt+dHW.WHi# >[I#A ӃyE7-iI2^ݭ$&h1z\H^\ݝPn vNyCZe ))F.6 HJ*r!Cr5EAHiVd q"WH;֮qi(W!(G`G z2@%GWSю\!3X%WkE"WHRKoFԞU"Ƿ&G4ܱsif (mfA WU/+QT iM]i ʕy*W_n_g}lcP8l!MCE$4^<"He6H9pNj/z!$W%#W+Lo!X.4J[S73PZI ^Y뀸Z*wBJr5E #;:/ד6 )#շ#WzϪs!|PӀ.DܑhHQ̖f,WOz)KJAEV )g\))&=TXiGFw-hQ:r5ARKKi0`pT i]R`pre@2HxDWDWHlrQ\MPaR`]$#WL"wBJ$:7=Ǖ ,(N9SaG~;nڹ>FmI(7t9x ?Jh#zu`n{wV8|__|AaCejX־nn{>z%e /'WgM}i[h=E,Z-6E包Bߤ?ÿ7g?ӿAoӲ} /5l>i{_M[w* NX.mvҟpחWak}6WKh|?J?I7}[r=vǷw_{vxSLY^eY6Q5RyEi ]%l>4*RwoJZ6rgt{*??K-d]f. ௛RYheU:)ka&D[!KFUMm.&JUea_& mzqWϋKhjmn57ebw?Z[Gbczz^{|go QOכXbWïj=s]^@O6m 䶋,s}"{ܤ>Pݖu}Wc/(xh޽sC^|bLeytp+KM6[r6mY%t~WkNC42/?̷ * 2 7!T#TS9~XW6V>IzLվec=^]"Dwx^.ۋ=N 6V6D L[W)oM;ڶuJu TW7`Li-!Q:SQmt+C#\-BYv"y h8_[EUSiT?mXUnjwl=_h D<'w3gCgteͱBᄄZuekX_ЦmYU+h]Qyib]Tnዲ}͊Um3fa`aAt 2iQV/uum(i:Y ӮnJ:@<"b4AEﵲpjl b%*+iMZue, GmiT+{V D5nTjӾ-n/Bwgu \9cINRd KA HȾ%XɢY#ɳg俧ؒ|ZϨ%vǯ‹I;PAC`:@l!*rOPB1eDU^Z<}gX},섞 yD >D3s ȊW!䬅@ *WRLQ07!OitWo6ёO-ۓ Wheª~tiI^Nod;:a=C#C $R۳;I ('Ro9%_yao\Gu1tk "ZapYy5I6piy}ӭ-N>on{sEʧ'C:9ڜz_txbrNyn6ZMrDsgm Kizpz1-%|rAx# ]D뼜ƪY3&J_Xΐ{9C^uxϺ݌[÷okg''??0pLS*2HVhQ9"x5h$':o5DO )z]@i`ÒWg/DStZ[=:}Vҧ) #/ng GPߎ'MүS++3{J" ~_~818 ȩ 2d_]=bC-C%B|fEUF1{2賩z[NeXLD VHT5<Z4g'z$u.Z,˙T$9btgMj/eo& =Z!~hdEr~Ȼ#noCwߢm;.]\1,CA" ƵJxDZ@N{d*鮶"@N@Ĭ (O l 1B6.PsZ*VdñoHElG(Y\t=T9$X X)C"dSn~ECa 궡VcS>"!.W)k^rlP#`V h݃tXR>ۮF,n2*D{TP[kTS MQV%J:ybv?T޵e.Н]lE MW jT .7qŇ]:l[3 4@iHˠӲw~0 x- hp€ϒبkw [ --}=ZKZ덄N|\ڸ@:o*8cN+U<ˆ2%P:brUŇP7K!(0Hl&4*)ڠxj]PJ(!d|g&S d M(Ldd^0PAWQT!z/ Xa,S]^hK\bd`"R6Ilm{=lQh 1 Iʥ2ѿ.cFKWZ Z׾ØwOxf ̽V`.4Oŗ ,_N2qﹾ?_^&A{v;feR &jg>JJH&E,W|}MK>]m'na!!Ͼ nzی_'d㻿Ej n y;$W[,oe:eWT~tʘu2&Pǒ"LQTeXكn-urB.hbtViJ*8M jrpXjLP+P(%PꍌyD%{7 isvB;a[!_4]7fOT+*SH0NMnCAn\DΈ"NxƗ(l %pm$ 8V@VaD(oǭj-7fVQqV]ÙTRF!g %ԐMʽ󈺄,: ..@)MK3. '\O)8B(NaBAT )TK6x  fY .iǡj5Fk7ܵ~F1&=>" n~|G@㾲f3t(ğY>uH{*-@jȂZ)q?iUw_)D.BnJDy8[e({|zE:LjMg -X[sRaB6EQUE:&/i) Sݬoh{Gʩ/}z\bHɂ'E*xoK&Q%T]EKBw#Wנ'>“hO>B;0xFPnP:5U&j,RⱈZ)^۩b;w0Mu)TJl-VHXpmbя7Ca"֛y5rLRT*-7< oMGjRCߥѡ/'U$%\iRB'i*wdMI L+tcR˪F& )ǢC*8j1TC6D OrX0V`rWVuںb0!-cq%;s fE5=N$UɢƋ"wQlL4&R ;8ćJ/neU {|%THW_Bd2ʣ|RǨP$q6z#CͻuM4 N@WJKn-я#Gb-=Ndi%dZ.vj-y[dCۜnp~v¿h5c 8A@V?>GO|w /!3ٻ6cWb%~a*G#W"e'š  %O^p#HJt|Y5-u1IQO |>rX_{@6ḛ,C4@֨'3J$dj* "+2)Q޽<%6>d:S-M8 pai,{p1EѫK, ?jJ óo6B\pzV|usID.%m-IfW~T ns1Yls_hȡuyza:;{^:js,^-shY' \˦ΛǍyUe{7O\nwˏc^W4?%yqu,C3^r*7bzuUOmvA}pn?`i{$OX=m C Tbm{p%HdF{3Ap2RY0HxwtmF)ۣl/\nI1 [4婲u"FjzI7 g :3:=VQl=4 ZT{gq)a@hNfAcr"gF^ާ kW{5LX`'[_E =!]Ta;GL'n'Ɠ W jJZ|OٳVode6KWMKn"'Jќdsk,QV8T]5?/ҁ %>eg6 7qDH8AKy0Ig{z @eEX\Irfm_gA'G%/>퀢my^xV3%fGɈe 4(S_caͣ:R"{D@'׿sPْCA£vŶ``_m|!#AL(C[M-OO"g4`T CsKFD͏s#?ޕ φwgK p)s1t"%@ф(q@ȧP)bԃQŽ|6/ȆK y Қ)KI S"!&"Ŝ(aFXs?>`չ]`׹jWv x4R3w_ ǯ|w%9X:# DzoR{]QI8`*k=h‰@,6dr\[ra4-W }q=rp PKucPbdEeRЀ08H6\g`kec&sv^u[qw`shݭy34}i 6pɉ,@ȪT,ԡZe!00l%&(h`ĩnTZm>8ޏE `9{'E*N Fg2#c:87 $' l}WXcpUeGs-g69C!R#Pp>D1X@i ,bijc%lx,<Fy*Ma9ҦTҞ=ǚu12}76( M H0 #8 E+Lp]B }׆a+'x>;3v}l:έGwD@4:&(fEwZI%HD0 AP$@2qU=T JB''5m>zU؟r`@,wɐ1;Im0HF=*XՈܑ_:5jkZEQF ɥ8* P%RFR2waAZfs`$AjlD@hL;1 sH"xJw;Q?jf|MjʶCֵ#[#OC)a?87/] DM'9ْ'Cf?C?&ɾ15J0;OA[srP?)|.,sȣf& SHND0: S3kr:> mW gE-jkrTa>Tp9" i;p(t| 3ucLHY}QUDG9!3DҡЩ[tYPç+_T9LϦTɽn'/oW.1JaJF\ϭNl>\sKxn'$j77âŕ&:/BzlOzQ7dq7en&nRVRO`g5\28^ Zk=E.!?2:>]zW̒' [cR _˗:̵zWw.~x*~?wˋ\\{vo` IF%e$N{]Ӧꮩbt͛u5~cW-7[ }ˎ5pEjek)C:1+@6udzEA_g; Q! 𛵨HV=~ܒzvQp\ Fѩ 2@(PE#a, 6sCBzmA` x\`>T Xa>M%PacÓ'>s$SP!;4r3Ûކ[߇)_Z0}F=1?hs^Ξ :́UНG Qv[q@`ܒKb|(do%*8ޒ{4L.&9:uPUV}WW@%GKOG] AKL]]=`IY]=\EW]=ZG]=J~+uŎjݥ`(u@ `U"CQWZ*QQ]=CuE(bd;6RW@D|UzR"< u+z8`"CQWZ]]%*|Q\E]%j)wuxTWG]1*јADN50AtDͰ?S穉)ص I&.\"ͼQ6;BVI֧<+c]ƙܠULmBSNS^ϔ9o[6P{ξmTBYr֒ik CmAڏ?r(6^NB'Iz_Q1')罘-_UET@n~+ ^&Iٞ)f#U *?!~"~')c)ξP%d_2|ߝfJb_@S%M0|;öڊiWbšeWk]v!|4[?ÉtU(v%3o݂NJAy}Hoxvk0Lw(&n)/8HBc@0Jʃ1ZJ񾣱D:#ɉVs؛XEpq3@Drnʜ 2#)np09sH[G‰uXT~_O5ZNqXt(>LU((@آUWg,3nPGuvvxY 3H(򹖜N)bxЄ/cfKum21a׎JKϷ̇ fk'}jIQN)\gcE/c^7`tM*M52A|.kɵ魨!'^;kU]j.j4gA޴}aa\ߘ^Y&YlT6F6;EKjfu[l;6Gp#UH㝑/[IQfv3E%C {MOjhUx85XP>=OfTJ&\^tzv>5r)iU ]w=W^nb1WۓbA4Ⱉ9|lzWU>ݬrTgOvõJ_9z^ 4qks;lz}ތG[W &\w;yOx#oně,Ј!"<"+3g̠\Er!u3}Twٔ3rdЅfޕ@+(H3vn#Lld5W9ϳuEQ*e眦7#}˦ u M䘬9+4U"?{Gr]@7kzzXa [ !]1;3+P;bw 93=ݧn{nwn] ]1oVNI1ЕлW@_6 0+zگuݹ=c^;Z]1`oCW 7MPF!]畳 +b p2K+Fk<]1Jk!]YZeI 7P&-9d5-+K~)th]xt(ztmÒ T nU{t:]=>y{CL\tRf㓓jo⸹Yo޼7o샂:ھO;iۭ8/ǎ [&m0d;xɿɦ\NM#BTݼErsX=mU0zC=T>8p&n>W1긟&.>1Uxr{vx'^Z?!i\>|}Oȩo܊V%Ç 'G Yw9|m~DL§8E 4|6y p ͑|wk{n|@/^o'zրol?*gCqĚ5.}-;LѾ$Vz7|W-#P9y\]!|N!+{_:mcs>>m5\ތRI8Jn i!ZOrKFn.Eͭzc+! حj.ԹZrS+ fjfXl4?8F:>ُ hX_[ЇjTIMҩ` ./pGDk3%2&6.M݃66YY2dClt)CQiԎƬr61b5)xٻ0F>8:LAZ XԠDtP 6C Zb@$F=(aL! #A KwA=[t{^,KT3f1*Tei5 !pBK.}?عCgu=y7<ܖ׵A [ŬΎ#D m3bka&a=Cw /M Td6Q t56b ڸLcuLC=GN`]@EE ]ʃJ I"92e(X LR|xLPBb2$kuk+w<o 1t9¬JrC#QhzBb9¶l^ A; > $dӍ??ߜuͻQZ|MC@;3 5\oF,*ga1=TV( Aw8+B A1-'jjZ e"~Cgj>C:kϢ;;kFFP[WCDЎ/5{%vLL7k! IHu |4F9KP0\qjh}0{y>,|On?pR+͂ ]@H77kF֞5`sΪ4QG]GjZkIQ3󬑲FhQAi7 @zVF6pPaRDHZN!F .`gr5wn+2TJ`]PLMYځ DGCPCzπuz\ ;fc_XA|۝bDuN M~#g=CE7 F-)Ba E5FHKCH4pY/xp`\Sclci0Ih j hN\76xk+fnQ4 C5kփ*H]6 |tf҃d &S@rZx tmKzg!|*6LLOfs jX*C*mP zxi ֚&-m(-\j䋑Q+Bj{@S AO8IZ{m(=mGe(utqH][E4bt~֘-jUetiL  kȎEЬ^UlCl.Xvk5c4&TsE}#W~fc2aԋAtlnrp/t@ڮ1f )ceֶV T,5Bqo}XBQ_Dm kPAHo}n7/n^]pvMϡAS ͏]^]t_[?9Fm=\m0~j $'* Q#H <)!–Jkb\֌'J }@`{L'H|@$> H|@$> H|@$> H|@$> H|@$> H|@QCՒȀ_ZZђ}> Fidt$> H|@$> H|@$> H|@$> H|@$> H|@$> =[!Kp74=yGz> dlo$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@z> ]zQ5f9> uy}@@y@D|F|@$> H|@$> H|@$> H|@$> H|@$> H|@$>ha=$޹>V/?奦~{i~9ݵv[u_`qgȧl>ll1s*lMy^lrW 7:bs__#Zք+Drm?~88Ե7Dc-*}p6\\V3}W~vǍw2Ў(]XSUi fNa8O  |B}$]p[hJq=GC֫R4_&_Xc 5_#D~{K mЮHSrG)ۗnS ڬ^)H _%7v;>n+ No Em"Ϸһ\z76>Z\ڇu=??<#Kz#i1R\Jz5-O=0ʠ%< -)-p\].)bw0zʧv?c^]\]{ȍ%"4aHټ_ yXLɾt`Y7+%Vh.."ѥH~ux,~t]ֳuRJR.}[ AEWHTBJS)jXQ 9Sdt\S WR J9<s q+ռt]!UWԕ JhUfZ_ul6 Szn޿yN/\94]Qz Ekb%49%5%ߢ]Fala꒦g,iEM>ZͻmzD$9'N0S?T_t]sGY#li|}}RY5?{jŮ:\k qP>)-8qi1J%[# .=ccZr;$46C`ɌWWuHdMplf3)J$6גI%!+>U]MPWSL+:7W*BZ+KRjr{RB`Igg iMBJ&+䔦A!+dA! )jtez{Ƽ~YХFUyMLiy),UWOzptޒJ% iR6kUt%5Z;EFWDW@k,]WHɫ+Ÿ( y`q+̖+jafFFW9]ct]!e4u>XU'CNH|:~_n^/ާoސU$۱'LU${&;tid1tiVu<w%a{,r1o}Z_ξ]%n=o/}n͢g6{Co7NOvh:ۗ-{7.ؾTehBXӊ9Y[Oo|&S7cNg4[\:(1!Npg9r^c<_5I"#scyNe p-sSCHNFWKg. Қuǧ+Ǚn"dt5z*BZWj\Ii "Tt eYq렲{.IyyK/2=4;NkNiP2♴;dHi}2&eg )]!؏Tt# IDue72EFWɤVuFWSԕzFHWG\:7 )K{"Cի{I`Wݍҗ+TUW_đU/ΘזL<`9pѾCʣK`++QuԪ*o PdtRP*U'T]B2BBZKEWHEJXuR[IFW$]!/~0uTRtZZ2B\e iMR:^u5E]ibNe?r]L5K4h( uCIeU{IwʫνP·˸+)u0;cIwf9.;܆i(ݨC`Ggl DƋ!ul6erJ$VW3*BZ#KRڪ)qIeXH: W9*BZKR_u5A]y.^|IIuJW++@Je]#^1٫n9#*W]&Qt%3t%Z;^$+5um.\u*<<ܱGkro٪J #+0NEWHMBJ᪮&+bp2B\+ i}PҶz]aKBB`k q+-ct]!%UWSԕ6FOJe(h'|S{ڊ[݌ /Ɍbr['jq~о=@%|p۝?p<5VzӲ;כͥI;ӻ.'zo߾?/Wn?G{hfk+ P??-o=aGXn]ڔ| |R<5s*徺^um2]]8(w{ywny,.;mKaM'z!5Sz]w.MgB\IW mz/%H6REyX_eqEzcVZњa҂<"g}L9y?g˜F(CҎe}c6 "b~'16xqj˄Gbr~Ira/UϘ]ܮ }?}x1M3 s/AR¿}_A ެgCFI2 TH:'MCo( eyrhTۦml8]{GJӀ424 h̸[T(2c2=\cz)Wçomx]/kjtMhіd0ʒ08:)vI0M R%Z/m3` ݘt[xm.)dc<^:`$z{E:Μ 'DEN]4!2n@.D+tB6ص=heDg9:pQO D}&p}˙ &eJxmoK aMaDtkwCZ70o@[AfuZ^_݄^߰ƴC]]uI<ݕKW3`^s陚)1̓'Z(1=_hk!Y'I4/}3ݴɅ5J+\*k,cZ?ΠlȤϑY3GW}^*)'zVv%{^N+ cvk=2(]2uh޺˾1x"zI$c3*Za.J=uG[n!kOԜ AB*f̱ks<!|p?(_n.K??Zv@bs(*u˴f*0\/g79\va|xغ=v [{?иE:z"D (K] 㘁Gy*&}ϵet3nYǃ!ڡ`P"AKXdwz2-J<ol@}L75|Ӑ@9~~}ݏS"~S7ú q8R=2g>i0L=[xcpcák҃L䈑 3\LzٻFrWtor{G,H~,#K^Y =Ŗd[a==3]ux]r8RhoN-Ue[MCw ?PWzSթfxp.0 IJ`AfR){*oؘb3NzAfr11@:H$H|DLaH8X1ύ R RdQsRO#}{HoZbbNfe-[];k.mVqSӛ-3iJ|S΁,Q$P8r0R \<g0;*Mk9X9mF9?( |4Oyb6 ޏ?|v9AIiҜ<_y{]mD@3}KBjnf _DOYb]1˜#)-?` n'ˠnI]ד.۟N.g.g7nmxw?Gw7Y2WB~ŊSebO`q?eR:Uۆ\6?$!}{qRͺH9CٽSu=B' 5G.R;]ucf1wv7ׁgZ-_￵T(*(q͙\`i?ĔdfMQ)kW\oZ!FPZ$#ci4eEQ}qw)46#ۂl-gyӥ7$;]?^<\3yT:u]䗱n!!4wfW@w=!ֿj#noEO:e*cbZ"z'a|> !m6,wNM=cns;]+||B=Չ')LǍOiL}Mj`tN5+S @_MjյydDM@N%+ԥ4W1-oO\y NwwXwGzzMkhE<]oFޝK@^R)yJEvyъk06FJpV7ܔl4#`:n*ݷ|݊5eڃw  O&7F.z%KQi]^OWrw62bXVWaܲU *I3DH~iuž$)Ks1'UfBݖ~,UBgA&F 57dRr)iKP_b{S>b&s#w;+?]"a+&΁ wBv; .Yԥ^} r< *IbN1-yLDiI*`V{ Ieo˿br"! pU0 8T͜hiwn~۫ȊjzM5eӇow\EWFak[9+ԮU{K/UQ6p/dnIO+cfD)L o ޤ0[XB]k5Lr`̒ 021[/KsGJs- \&ُJ5,br“H+n.xKr`W&x\{ۆŤ>7>O#6㙰FH`l"R4}L$].d #n`2MCq\'6.5’KCV2 +#v5q#M/]M;ÌE9spŝl/1J$7V9xu}qG!3@^tĂI`D RBEȩNΗ&~Er.?$o\{A=p'.+0Ȑ'sWNh^ҼT]TpcQՍ@ y1΋eZ(=1?>9i.؋ڷY%fA8d&tF$G\1"){˾Z,Q0.#խ.BJ傴.E&k3p!k s!  vjg)QO IGl7=ʑxs~}57}vEp>E},'N۳ Wz-7A#Y^g hOjsDuDu޻DuޣDuދDuDLFzPtYVeO$(= 6JKva~`UX !t}"܅ [۔_ooog Z/پ$|i'h$YI;i!pnmyBRz Azy,eDs+">ɕbU!Ȍ)UDKޘ}`1 TnoQ!Sl>m㇏նG9S+mҦlޏ4>Qy9"~u-.qh`-o&ae!!}\F'PO!2~힖D˗xvtEvtE+ܸW(4|4rEeJ"jze&c][cdV.o8 I@ui@#hdQ$!Xグ-RE1MAm=0O^MmBM%^(1H*U\wSUmk_\;fX5IaцW}!UB̃!j I O#ށQCD @,͍Pdjusҙ%L!M( !h$GbN9'%*vvɩ@r'ᾋ ,o[o{y!guZ&1,@c2$ X/Du4*S?8lKܲd⠸eqw5}wiiҺ{둌.N_兯%,7YeK1ׯO!@֎R xVQ~+oU¦⥟d8;'`Wx\}y(EtJB8.JdpDKjqVNp&fɳd{Cm2>^ȓ-mVg jep@{\ܰpF[tsu|E9yw~)#. Xt̯l<+$pQY755Jc?D yypgwLo~hnc4E!Z.RG iP$E_¢==~ayBN"蜤W٠ #uS";b@lD:A3P)ZBa YmͣC㐦`IВMI(^M5\ q^_nIgɽr.7# ZiU 8* J_D}RI',D<7NXf9tȵi{aw6߹ )R**)MIhE-(9tw+u<wC8VZ^hԮ =XJZ R $&"z ^PCRLc$X/-Q B RqQe顦}2F0p@YŬҊ NY:a9,6oH r{ K53n$LD)<[H!JͨSIEvtc쟹>9[*?sҐ;-Wp%8[啀8L,hbӈL2T/U;?`]Z4=~}԰S.$9ȷcoljx_f,8fyzp h 5d(Krr$; 7r2# ^đ){QVd%ojDg#1ҁD.pl?+gYxS*/[a`gc2%ԭ-bWNOWJEbى),rmU(U̴j2MMVT)9Blԩ~zP=wrΛ˫«u[0ZSΖ`.8fնk $(j㟮eTWxٸ8~!UZ$`{Ia1fg1bIQp}.w2}Y4lUb8r2rsJ*uG]",{{&S?}5CJqr>*Q&~>6W~x/ُ^1?(gݛ8[`=N ';!ȣ!zJM;͛*[͵اhhuSn;vҠZ ~r^g>'p(7ij֬|c޵0W6jOd8%]C*Y|"Q8LhuFiShѡcb}vӞ ^h))yZj$r"׊:2RK|aAyE9zq徆G4h|F:K Z OȫkViCQLעid_UD} J/ 7ժq< ϴ:|] G9Y@pBR QJC*};gu)ġe821I6&E#!!G!UMITeć TL$ZjRjl"*GAYy]MBNNO;??S=˹yIbamF዇>?cB.o<;Í UhR!T dFbEMA=$4' Z{k%xյ)oGnY2fmhyC跽}H Păv(u:5a$ DqZ0NB:+( kE6I>R|QI# wU gķծvcmnkͨ>jaqh|>7YFJ-tiKQx,{f PQ%#GdIPHN;kd4Jc,X ohkԡVt*ŐO$8'GV=;9Zz䰼>0@ vEoY>e) qr3r vY..G@ J90v\[Y8MFb4ۖc[ҖG;ߢߴOXQxP6rd/3$+4P+ "9aUD( L_>mضG^8,KCC?[nDl6Ym}~4!} FqJʇ! <<,rw&i8V>E΂7`\j]r}y}_2&߼ϪN-Or] uf) v}UsCY/5U/2ӕ`m/QK>v~*V3tjh  Ğ]+Wt`N#>vu7?] %vѕ]z gCt pAw2ZiNWR=]=Ab+, ]ew2ZmxOWOPZt0hW.ЮUF+i*T'HWBR]2U:c fr<]eBt t2`):CWvwDٯ]=I0rZq5efv1 Ber>+qԙ &3^d62^3?p`x29!\F-˔*_r&<K>h=Tto_xƝ$Rp]!!%ڦ*ŴhBxQ̧tĮ}%^$vW-;W$Y!8 Q 5 ^xs?{q5ivY\.SItZ $ R#BXRD!{)BHYނPv%\'8߾h4 )'nZU_^-_Pex6b6L2R W).ʚ$ *4QUZTo/οl/\QpuJrw%'лYvv aRKym}iy\=e"u OLLh$- tTj,ɉqI$SB*%׹__kH ]C;N'>؜y[2|1yWXɒC FDBBg,pR8Z<fDFdkuP:8eNP9 )MH@$ (\`>p3@ER$H-]Mq$P/gVhJI*F#J(D;MM{3 "RF%sFր1-6pqWA1jTd)%R% %K2PH$cpLhqwt 3vcbLRfT1xim,PCm O6;kTA|H4EaFzÎ*ӠPqUXwIq.Gib0˨s ߧʷ@PRk#v4*6U&)=Y'蘲xhMQF"# ?{׶Ǒd]씔c_y؂'I%X!=HlvW̌_7/eMV1\΢z@֩ X2|,R $J!wӧAgUA2țk#8s#ԬF[ ރF%k >Úݜ|OZ#vR=cGZG7H}m fQU>6t&$7'XK!Q,KBJcuߺq)hK -)X'5/V[ZQW!5*dlzaCKU+ܜ-Օ/H!xɠݳ. R^Q,f^$L'HH/>XԠA89l,@'@Hu{PH&C@U+cG=D &23.}$ ^ܣE{׋qJb6(!9Yh>3(RAUvNZ'$̿2COvיxaow˚S~}ZN [ŬΎ#D m3bka&a=Cw /%y:d6Q t56b ڸLcuLC=GNf .' ) >@(E&rZ#d^S1P>fbw gt,Q~^%f0A 9ɐQgk#ԭHw 3xh҇ *YȩՏoTG}^ yT'ێjJz%B4Da&ALv|l//]D]eb:\yS0x6C@FD4=R\;mŬuԆHT*Qw}TPGQpu $$7(X1z, mM!gT]lJvl 䁈:u)C`E[LnV2ÈыA(Ph0Q! td+!x݆`m BgViJ)C2~ȃ 8vGyQwEa2,Ƣ cA8gPH!(DMRuPkV]Ŀ?yWН5MVU#YA jfᭇCi*hhǷޚདྷPE;L7k! IHu |4|Vs`o ֛.rl8X?y7=t/`|9o^_lzL'A.nn8c`3 =k(؇kgUCŨծ[s5|5ǤͨyHY#h4v˨ ƴeFKo=H#t5n(!/{úSi樇*tyF0C{s":Y'%\aۊLv0X$SSAv =>`yÎWVhn"$uSjCn ?XwtH.P ê Q GʅPeQERPGq1*11<RuN` tB)1QFǤz54 hc3(:C5kփ*H]6 |tg҃d &S@rZx tmKpz] %7\sA-4p^ kJ> VAP4kMyQ6.5(}!Z)AH8IZ{m(=mGm(utqH4JC`Yc6\WAҘ"&b9R Y 5ɦ7")SJ,;wR@ǵ1[ Y\*y׹]O|w"BPQ>Jb`j?&nm^VܲnL2v*äIfZ)RuW%uL 2 kPAHo}ngwn ޙKl76Z߾9nѮsˏMsDEWݯ͗ywlwҏϵ MpvV|\_78ߺChۭu׾//}[v?z'CՁAOk_ϓn\]Ѧ?)('坤$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@(WఞCRntmZy@^|@>$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@:UP|@z|@ZeVZgX|@$H|@$> H|@$> H|@$> H|@$> H|@$> H|@NQnM> v+:/!.1}@w$> H|@$> H|@$> H|@$> H|@$> H|@$> 轍Nэ}so5V7[oӿ ޮovo7k-pZE DƶV-޶Ŷt  aEtVa-thOW ULt8$va5tp}\ ]1hNW28RFYtu`瞘qh}:e\X1菠+/tCMjMtŀ)k+Fj2 ] ]!G ++jFb),2=X:*&uj pxu()i_]1u^5tb ] ]9mZ]1`Z]1\BW֧JyeݪoU+_1z[U{ڻDZf{@[:nq?\mwꏽѾ'@aд{r=]z}הW}MZ%_MO#[# w/-zDcy1:޳6P|q2*}8_* ˋׯw3v[|Nf_oԂӵin&M)8j&A. ?f?RSE1;!NDYMq|H]zp%+ȲQ:&km('?ΣUCZ2~>|Ќ}q`֭~6;{MobTm;w\Ϳ)|ׯyݼa.q֓Ie3?,?^/xlmƀ]MmN6cf۠5R`mtpqMt_\tpjvm2Z$FI *RZ]EYϾkZђ[:]1>ӡd0?٘j:?6,e} 8⓰ԓQ~굫'޵yZDq(Šp]zޥE7A?׫.(b-xrBW'HWƄDnEte zjb/BW'HW֤֤UʪMQW@C\:]Y>>] ]5DWLj=tpS\ ]-ꬥPz#Y~MKzAkV[<]1JQWIWΥHcbh31 ߽4]!\~h/G70o,ʉ;3zDUaD 2% *=؅$3 >'7*!Ν{fZq=dFv{}+RȬFe0\%gFT<e$Q'2MڬpXOQp^ ]1DK+] ]E\]1`O+VmZ:JdRXӒs"TX ]1\;d.^]1J;d8"rKS>Ц'zQ(#-l*AWQs^_]{Åԇ ֤IP (ùPR_J(ZMtJ+/BW֩HfEtɯʤMftzikWBW_Ymӊ ͓od8 Y?:xVNxww4>ƄrL rrjYluZMVdv5kvkfdSȊ߼7 -l&u|`SdjSݏ#+|U v}!hUsʎ:Zg(g%-3 񸺣9ǟI82o/_/w;8ۋg۟;.wo_|q|m#ח=9]i;Ɔ5p_t>_{u4||ߏ ~9ݏ|aLj'3DTLj0}q2AxD:=<٨pn9gtoE:v{+0gK uho0()b.?{gF"MneGy br3o.@jؖW[wlRiYj`3ڭYVIVrEJ *\ej\GӍ> %fJzt ~ʛ|1n}qR;> o[U|?nя(}J챓6nƯ7_K[6_A7=w,aec::?o7ӫ[ Ϊ#@N/UyZ*UQypOL|B+?fQTGWv4jmx~=x؞Yz%Xt[%BZ'"`g:6SaDDrUS]/GBKӇKD[q8ZH(ubSa!BB^ZU`$Wlc#WBZDqAg\)' &y.B\\O>BJmL8Ph6rEs+\rGNis:F6+&]mMer Q0'1Nƛ;(10[Yi n D#DHkAe F)1G)\rJj.r LrE?rurFrE22ZRJYz(W{#G$DkOG}-\p qb]-(S;rJ0'㹫HmtG:i?R\r FrE"WHHJ\PTP^{FrEް+#WHK]2Y#W䊀-p"WDdrY#W:8a#"GW EV&]rG26K]90]N kZ =nAȃ5M8#3.F>N8j %]H.fs\QĢ+!W2SM/a$Wl"\vtK"UJI $QQtE|+"u"Yz(W AKH"q iRQrCH)B`\q\h]sWD\Qt4#"`'$z\ebgg:\ROk {aV4]'|IŸ/tII3)nfJHE MAMRDٍLe'p/))"J\hI]R,W=+NSlL9g3F%B(!?\e$We`#Wج?BpX{ ˑ+j$T lF]Ux\(]RYjz "+<ulhڌC),W=+JyH:? \mW'Qv2A6w^O*(QVBrEJf\i*(Q \4\hӏҘ,W=+YH{>S"WhrR,W}+cPI]kL_2uZdRK֥rA D6bl+GEY7e)18'#kV} v*LvmknQnf:ƙϕ!KnF]&/9Q{+W"ܮDZS+yfk\lp"WH@Ru}qOW >/WYE2(9z9r-M#aS|OLn{bh}G[(CbrrYjzi@yNr\5\hJ]2,W=+eВ\!T\QKEqe\q9ki*EBrE:UJ3+v|p=E"d(S{3(retMhn:/X%}hKJm LmsDچ<gu6(P~ d*FrJiع}zR4׽ys[pU~Z_헂&X 0(\=nLh" VE(6yE=Ӎ_@"(YԻ l\o]A׍y*}OK-qzʢ3z@|7^O66r^56 ._I tSFU- 1Bjܫ%v :ɮv{5dmu= \ڞ$-xmw)iM'[wt=IcyPGg^U{|QгZ~U y*5$%-UdžRz$}'3zBLV)}gv)ֺ!k9%OJbclJ͛|'rHo+nT7k4`b@i~rOB) xZ,/C"AyzF`^sדiaA5XTg LE9.Ycѡ$r4p cScf4B{zOw'3Z}g\띞D@gx5/sJDyek'[QvQ*8p5.?a8Y? 狓?\dpF<5/f8Y|&h-ԘS gqY`练jkT *ո""6iNo}4R=xߏRPZWp竢\?E1zaպ,J;])^UN^rt @`ⷄFh]On+iQ{ !pt{RD5l"K~)"Q3z0ӈ+"WDdrE \P:Fr8p5mDkK\-MGɐ@`B'Vꎎ4R#Jgz饃\IeUn\ iա[_R+-(rUZ0+GWV%]%,W=+p.HX7p+2u"J>ʕvܕBt~\\h.WD!U6;Xg<1P1:q-ťθny5_/LeY)r+ǙcQHO3.@SO)5`gG;E.F[&7#\gfDmQsfփц\#W끋\mPRZQ7NIFrE\Q\hm e\y:ޓStNH6rUYW@~\%u}}Њ\0+ulAQ,W/GLKӛd{b;3;^Gk:J#)CZre"dzePZ9FrEF+µ\wQ媇rStEp"WD뒗+l,r`䊀;\!Q*媇r?>\!;\rJ.rA(rC2ƴaaۗv?6A o&ZJmBq6՞EG (ϱ vیbff,JYKbZM\埮\X(\q)wNe #PZQڷ `eX!EpeJ Qv S2e8Cq="ZQ媇r R1+T1\hO]ҙ,W}+ `8!#`/Z6r!)gBe\x9z*uy;jw]GMj%vWn\r*!8TBGV6|pOEs^n_Y(hbzoIɖ-ŧ@XO*ua`d2+mL] G4Füvyzhp?T.EUWԍ7SDzاuJTl1Dݔpgc63 ث!Nc4dZ7b:'+ggoRmou{u?zܪFbVKY0 MQġv_^ίd^{*r*o?Wٗ7"+^ӟ7/lܺ?0ѻX./j7_~70ǿi\է`ʲvV1U6*B9(g*UA[h} )P6qńyqs=E/Q{=#Cu;C6o#ialMG[|'#Z6+G?Ri)Plh[юh(݌o5wrst{9}@{ `PRqb9>UW/V?]7j5(vROTT){u5*A=CXki!ƳZ(@pgPDɆ=VTFyھ]ƘċŴˊFdת$ږ*hnU)ed869Ų|QTes&\?3bhqVyg!0-_lh GX>K%L_VXaܬ#կ Z^OV^yr~w6f}r#ok+۷f&~yXx_ÿJ21R7Yg28kL ؉fҗ^˩0TcIKMΨX 7ģ1]/kT)Yሦ# Y/K/6q+'ddS쭋Mle7y!R Å |[lnڈxu-]QdV~#r\˸n7"Dž-ד]MDr-?j3*Qvޅ|9|QFSn -(3ce놁ReƬ ڹ/}Ŷ˦"|15 \F8*_2 *ܼ@:?cz!e?7&~a6)yXƬ1KoO*tI[ݫ݈l\)f.HNV3;[ & sRL1fl:c){S1~? Mw-0ziy?bD=2.a6%b> X~A_{ +f_ޙ5Q~jKfEimPD5AA*NG>ۜ<{kZ㵣zZZטWĵTZn $|Qm˖E..AQD']~J s snz(}=Q-p`٘Fx̪[I.s&f39fM;2gk6'(p?\tzx$u?!xxq?<O_Z,/e)epP׏ fHl*hFUn>~yv5kӮgvDZBf` a μl)&N|+I̿/%"ےC,9.%E*iDE;.WPV esZHS,1BAjNE l4dDK(⩠O54;U:4@AWqw5;/uq~TKqA7V^*bFIR'j$:) w"(#,A+4N-ǎO1fX) jw>ZjS{o*\+n>T+}.yu!*j<)a%`H]3R[U5BQRvcAc)&j Woy#𜃞=7H']chζB2"h0v.z𓲄lẸg  3/t:F]ܳ&\b~%85J5/WqBm)M~QTcxU<ճx:+,/Eql$:kKkmgm —-,/-SƢ"EZ3m /&mZQhZS6SԳL_Z5Sg,ܥyY=+Ky'PռV@+2+A[ʭhS5#ʳܘ̉F$vȻ#$k1L(dNL/f9 wU- xX8vĄO'6wduH2=ڒcSm<+Q(- 1cK԰}䭀x4Fg \V1%t>+ij܊EuG,9YEz Ġ5񵃋  2,&*O.FAR9ls:/#Ҵ=k 1u`5}_&5Z0pG^F^}}ӄ0${:1s(cib9% ePqbZ!n;vW+3W(޺B$W_~fR3C3w,G @} /n}N+s E;1*1A3[ ?AsyukR8l b XV/G*YN _K*1ς]2l5ҟ ZC6?ix`n=RYWI?_ r{bWٗEz"8 u9vXlN&J7Gy9>qx(݆|CW9ZV#0jBC]i'H<`y7P,t9DP}ʹ!;bm /kuߌy*G8ыɵ2+kWLQBLl(AU/QZTnHl݊ɴ:+Ag {$]]"y8MSθl-tJhl}$)0;FS39KBNM tY+%7[/H9pn+ـq:>>hM%smmPZ;TϜ<&F?d>̖&i0\" ёġ`tr*qP_7pRnLs;]|ٶ aPы8}YcbAĕK9[HzomE,;}BYQܬ,8 YKuVq>\?aWvO(IQxyƂ@oSٵ6sM;hzso`NqV %̮ aXI{̎,0US CjRgRX#Z@ c;:0@NUdS˜@UtSk !q@B7@=-GPT$v# є:4P새g3 ♄@udgr@(F[jsp#,z6B0|ǃG!w]ÌKwd 2ۓcwd"Sx#N1[:*)٤?w!)D»Zku"o0W&OO!9"bh'm@7ЅB$5v~ (WkTG`i FOŖ+ý.U$0𦰊wA I·RW*Rt;0{j # ͝F99w1*CG sh88-[b>w4 O!ͻG04zyGO"?SX6$;>[TgK|LJ%RJ|*+GI[Bvk1.B1jL9ǛǰTs gޕasΉ. cI; Q ܮ,w(|!ILNz1cҤB,!Q/R>&AJyѓJƤLډ5i1 W< a/WdY|K\1G5S$w&8G=B\c"4 W*#d&? 8~9=W]3]ͻ; ƥ6-M Qt) 1UWo9Cr:7foPHHrb ۊnx9ẘ@p i ϫ۠U.%Q/o&cCBk }߽XB@\dfUUzqt* WT$0V! <:G ૜XlWEw QgU'MW!=C9""Յ`$걷Rw}Lp_Oae̿`T򰃲9΍J^~]$:q$"42bdn+:~;MQtT erG'BJÂ0(~3cRʜ(T~FU*[WZ]~,\k;;yUsU8< != ?Ue Y&\q IpuhJHh6*4& .L;Ӿ,ɐoxykǔȾ.ǫI<t_6@8{{Ow:U%$]%oOHMlRPw5Z2aț O"Do'&D~ziJD:o_')8 {Wt H}[:26݁iON.Ru󭩰62UT߿rYa!ZPQ?Gwx +7Pj)^RpNAi#+p)bXGȸIДñc;] w}g !XWX3RʘTǍxPԕD 'T6ME EĒƤęmKַ󻧠#sLLeݝ:'u].قI"3&|fv~o=o".q r4yPB4zyBdm/#vXb^"90ӹ7L?l5?}~lr3\Lf8RO_9AuonW_C_,ϗ_+gw~,9f[M1 h2)}Qa?yPqg|2rH۳v9`ǝXȬ4W>R̒5$K:E|Q ^}%$ʭ MlRKS}IȈڣ1*f)6crɜdI&ܪ45뿢 0j̡SaǢ=+άiam]6l/G@a8VuDĦ,pV9Ƕ;Y5Vod|͟J4#$$I(ݪBn0P Vf*(١:ajQdL5GAܓYR M*J@7(Q:#í^AÅdطxpT=s?ڱ"R5!PV!86PeJmn?a "RUjiͭxWO͔q*ƀ2Abx8(I {)*փ+ڍ!v  8sW3 Uy`c<@>{bt D[~#;yL&$GlԎ*N=fN6s 2j& (h` 1$M5-c&C֭xd |ʌ1e@$Bj2 ,^csDw+V&\δ' G39;?CэY7M )/cK H@*r֯ps}>SNvh=ךjf=QUh  Xw1VAy RB}k4c GdI⌾, 5 T!fFWA#ׁ >Sg8`NMw`wr tŒG;0P{a6&|Rm8O]fg-7>=9,Hȷ-r")Zل 0ZaA  ?S{ikSPQD:R~1dכtE 8}ԋ8?PS} {|F*4֝\T="h4N,N9,L {w#S}Jdn X&[?Leݳt ࠃKTgN:Yrh'=#qJ KxmFt^mR3!F=#plouͫ}Om>cCS_J^Һ/˟{BwJǫ]ʞۼKQQ)=RW[0/oXeOS}N8"蕲1櫫^[I3R}6+;©;FUad(Qm9ڿgՁ\ 1"?kh + 8]Z?Vw-j}88ʋ ה/]_esAGKڴyUyi;7)SUg}ذǧ%?a|n1G؋]6q.O{SnSʀ4bs pT{}ߥsחYa$r+hroı;̄0S_xŞYOY &cq>Lh% Y OhƔ1'|\\C|2ow}:xLuݚ_l1ILN)jm| Lpv¡=T>rDzT^4) 1we#I< vGE/Ӎ{aXUH*IJ>(LJa[*fȸա&TB3RxKW` qIRIx3[>n/NEߥy৊É6$ՁBP?y˧I?a6X;DˍKVxCU[é7hKo5 t(gq&w^E+1;`E__~}G/#vRDu(>b#NCNj.7 PLOJ?r8I_lTbOygj. İ1 =3ʤ]1EmUHɩ-oflHW'?F0gzJZ7DLRF HXRT#_Հp)Ow̄p#!V`lS.KynQ, .wN8*rߜ ?Q%W= 47pDM-S!D6gr=½lpn}R,7l°ԞǴox4"PNLgajCɁkfNH 'Nc-"~& $3@1,\[_Fǰqa,o_'$Eq?Gz$\22B@w$O&l,;W|׳)2?)" u*d⽄t0CC6&y;rpG-Xyeb%1@Z2!kߩ)(1<3V߿-ϴ_ E3sLB|}h\jD0 #T Yiu1wX [g匭's ގBԮa*=ZHi2|3RQ ׷^[)UB#X_j׺)vec8sŁ 25Dc˚)P䈜?H#C$cei :aN|:k֋i-auӟM3(xDMbe^?E.!_!*[; AoFOlVQc tNl1E*XcwCA{dVu1$+¥+[ p /A$QKO2=x(mvgfҎZYb|Zz=9mlkO1zSj%/]&N%uT>87tI,̫{Zuq@2p:ȎڗT SIJ]ܔפ-H)CwQUcWUPSo4-u{=T㹣tDQX%4:kJǁF.[*zY\")b{S$,tvqP:ϳpZc*!NYk7];: (q֕DDlP'0%R73}'<PY!#t8DAJQ$XgN$4{CSvҽRb]n4$q2K'ݾexH` >1PX㼯z@{`^cK"u:8}h^պsplߌ#\6rL+/n%:97\s}F O7b9{q\|0}o#նP`[c.0; ;R5q`8~Q2CFHg LtⰩ>E\btJ{bZ;wsJAH(i=9~N 9 ߵtP]Pog5_MJcTȈz1C .n…$+[?K:XM& JϢ ~-JNX^wؤW?I04DcX]f^ 1 j @lg^Su_p|DKprAP:A 8Fp @(.j|.|_EEH:^N^X2d;|rE~ ZkZeqYTJ*Ra| ʰstgqVrb!^*e>dN|6[OzxFsn7s__'d;Q“$W0M~Q&.y JKo~Y$)$ro~]]Eע&,*%d|uVFCIt:͗<jZw_ޅ;dB)N#&6)p} ٶخ߬!$P\b 癩{_x0IvE7UOF!#Hμ.!\tXԦؘfN-ͥFyN_6O4:-"S/qS/Q_%z-9};ph-CA8b;Z'm{;!Ғ,.NvZ,%q\+?FҋxRulf%{14zӫ67"WcIy}@VLc4QҬO X/U|~?$8μIV3 Xs|e!iQ fP֝7e #w]۰k&yY>bҬˎ٘+FpVWnF:B},^P(3oqifǚeRXZ{ޣN}.'B䤴I9^e!x ͫ߀ư 6f8o) 6Hi` @Ƈ DJ\hSx)?cY_}3mnGҥ-Χg ߊxUtfu[#zɔր׺ɼram37݂ L*;T=BB|q,X>i3kJGj> 46CEia!biWX?L ~ڭl);Ьwg 8(688mO"y6`Q(DliK!!Q ,`߃%??*B5 t-lTSgi`5TD3 (V t-c(N ?o[Vz >oA'[YP6]y=ߞO/uFczjdu7SlVZE0O+UMpa1o]vtS,c0°2R=i;w AOϽo<':.Q' pg IljbN DْXd+wmm$W~ٲ~1gv`,v&OF]$Y%$k0}OQռXf7 -Nu}u9=QWŏ3\KxMdgiwH 9k+wj$XT6[i>^_4*LV U=qǯ b1WXM8ΡOA\G$qeٟ61e̡Fa,[p405+d;w>2(nL#EZ!'5J*2 n w+>YmӛbT3:xozM(~֋]V!JI 笈*hKebe@#aAV*b׵Ehxc.#)H] A7VKiu}Whd5F;L[(:4Fލ E[!xuk`(x˘LdzIxN$^9P(kVNK ox aEqrh+du ,Ls_Fo;Ilθmu5P5^Ԭ2߲zBxf *4D*Ir*)-BVC=6zČpCjs[q:~ oBn ?)"{x3H]gLeoz y: ˧p'&s-N :k-Kj+͙i~>C8DI۩w'64aך9#up*d&RR߶NUsP [ٙWf5$XOs|{> ?rD>do~Vcy>? IISu $%m_]9+LY4Z A QOXa X$ SI$p F @NFcY3UQPC\p FfL`l]Vu%"  I )QppQ՗wboˍ*!қ#UagٳhpkM Y)n7p.<ï:h5OZt4ń[Q9i%BpHdƨ&'ߘ2 ZKFP^Gfz!CO,Lע>]Fֵ۟f̤jat~]>GQ \!5jcs2V^^q uv> ˖f`J6 ZUV+?D~\-(DT+;937A7v$` FnEkWaIT:"b*Ɏ4@T &3 a -?~I_R!T\Zs/^BP5F* ǑJeЧ*єXIvFqwK k݄9PLFqSJW|)%`RhQ.!"qhPE(jU! 4Qe{Ë0ooFs@OO1?>/ߦt['Vw$E#DR`i*"NGZj ,)R30>ۦ&)@1d)a"CF69ˑ ̥$,'f`[AbU9*#1bDž1hyܥԗXvlfշosW9R^^yVwRUC=Qxޅ'fpR_9[Mwy:͐9"lf`/U(XH`#[b:`9ԠqZnv%(-:SɱwFi!`y쮙PZ<1"%~.‘kjKjĕ~_iѣK7nΔN0f;PQ(F92z~H?`nmQq&YM=⠖ v=JNd <)%B ΡL^#Zِ;fu,Af޲/vT mC@3PAqMOƝ4hE&oIY#eIYP-F8wxC\ƔVLʹP^މ}ԎfwRLhNz-'/HdFMlrP݀By/dDwWYQ~ VgՂS;Zm>Poy#C Fb SH- T&\Z?i~Zy1%~LJ2,RMJ_ b|}9M,!,LQ5|1ʨ_0g_?G>Š>o@`|=ǧOr?UޚنʦmӬm>zaO9ا;޴Lf\RN>O._yFif$O1RZYipO_WA_>tRSug]jh #a|ێ9,z֒°[99c0ouQK%@ ײW8GfWx!,Gf#RIýAQtRŠ2Ѱ9aB"RMj/FImz>*rnttM3S=r g{onbw jp#r8U{M\#>E o˘JrSS8нFI9ׅUٍ]V?aIJ5N,8bdKwNcAcDi;iR߇C0N(ѽò*C+ dkϦ]zg˾lÙ_o&)7E\MU5Y!/wu9xnIʊ¼",],,6Y=é<ɵ5 6 %E8s(hKmDZ8t@*Zfi5,eZ+Qn#;PhaJx.F׷q.S\Np9+)@I=5asLO6|_Ƞ̪;kX"]xg!U}ehɎ?q:tg?7+f ~HCVJGQpbG mYwO6\Q=..5lfxxUwh2z! Y:;)l0>'.ecM;fcLyF"3 *c^]jnOutn׈j5 6"G"=% E1Z=L$~)6^&N7jtr?K.Rj'.rJm^RV"z^Rsz]Ax히,d(a#Ldatʵ4ؤZJx Ys 3g{#B>o 9_Un`zʵVs&DB{QSasoȤiX$Kx#ħ He}lNJy| Т|.˙:bgTxOW&K`.K\FmXpн)htbD Z%=N,NDvKMhxOٷצJ.sOC>h͐nk?D|Ndfh<2N PJ=Gkd4T5 eݪ=ILZf+߰d9LVhe9L E)3kq>w H){6tk{)L'4m>"Bj}3 R@B5Fh|#A)qq]}v3z:8G: w8_cy6ueAЖNhUٽ%z+/si`jtG&Qꑹka?܏X̜U_>*Ƌ۴'Wcnə2E[/NMI2]/w0cpRRT}s2:0* IއmX 0_Ť=&vΫgkoπZGh{]ىm'@L{:<(tx"kHh Z'v xEOy{ݳuz:}֨mNx̀Jh/}Ă*-cDyauNvmF,]]( ;1-~6SM b*24jdM]{*o=[,Iv= OT{wxN;R PČ0WJ(" $XN-8("dijQ ~@:2#QW'(73SMNM1]E+Od{ zUu9p?]M>dvgCeJdiطHgZS3N{q2Lۋa 鬀C ) \`¬A/2VJ=NrDJJ4&1pt^\ݬ4 o?2ȸN+yj ?^aFc$D[j+ċ wŧzJk(9Uv2AggeQhJ6ԶPu1S\!w>cEu*|G$m0dKBeڔ:"w}Vpe1Z(-U##^KHd DM]YoI+^ve睑0k{ z}`f<%ٴ(vFRW",V%QȌ;8nѹY;eh չIHC.bSVP<|!GԈg5φBF`vÁH%w 1'1 ͞2W@s`LX>LKf?FɢAތ~}?<xvH4;6c¨To/FJz!hq1K8Dh Hyݛчt21g9~L\"ق{ =LWi>Q;7;~?=F(l?NǶcTGC}t|24{߯:k!4Jy:d(pٵ4dŻ7еe7d7ek74WCCԪ/t C@t :un]GdtmaW0a 3ZSֵ3;7e݋yUj|k&sy.=xfyjcު|rI:cKv( q"pd}%W+zo,ۊCoM3"vdae ǩ$c,1sJj˔F5sԓg.}Sw$Jv/~ZaXb  ny^ʬUC+u 5WϠg蘻e`0oB襣dۄ 鮃.nmSO_2>0w; F> 0[~M{{j06`nxN.iyaF\Qg4qHuvVlX_ê`]62Řg(H) JXM+TRJ:I4!C2LpFQErdCM_ t)UQӇLųxMk?LF3{bL*oOF.^6ݾ/yI#zG $vw*V,*\!|u@NL 6>E:\DMc'˅dASiVbzO؅8+E2hyTxPJR5^Y^BPqC .-62Q%[ FiAb7kU^xA5}ZYMB$ҝ, k(ӝ b՝׿C!A:x M:Dvם u6Lܗqz4K "<5k"RAߥkB y'=7 ۥ36WF[NTL1#o֬q_^e  |~>duwl2oV-@WO2IP9y;\6y*.ꀂ82M|')uN~is޸5.n}/ x@T]Ï&\䊚o{J woKhh~1/j)-Ky0+ 㘇qH>;2~qAܲhΚ5^8'o䩙C9@׬G//)4+fW6pQ3 R2:jꝣD7vߣJZ$p?-||)ÑoծYcgqk9JͨSq@dYcϴ&KueɎG8Um$ p*O;L-/=4/[TQݚeVA1\ȇR/10h%bm1SzsפvdRݚ5^: l#$prV=oo+DwK/ R5%#:׬*FuQv=Xd1j/-B-Q^X)mn=J.M ߄#7+YM;j g WI%iDeZq m5%^qI -gV oXt^84wK-=J Sf8k ~l$Ȇެ0ռ=bͅ]x~4:h_ŹC8pKD0}˜s1C4D#7Y/)&P,*JkFd^39,L 5 7d:G^_~LJۿrav H$xfs!,)T4Vx씓IV{])kK͂yobC9pbQQ0났E9sɂ[m4)Ԋ #B1)aR]ٶَFñ3S:@(:Qr4|0^~I4(#$HZAPHA%ϚK``wmħuhR4pcMo2EPmqVΛ5MTuw[cP-]pct QA`Shٰ(EOjQr]񘑼}rΩd *뼠΢8/TY-qG\ph3횻j etOS5I!PU xӁ:U%x2HPNPZFehJk\9f!Ô[uШiC OHI=! IiJ!,zǛถm* ho'h`"JwGǰdJ)ES]bHcD8>i;hy.GD>tJ MKɂR>eJoLӛmX၁NO̦KI*+㭯hG)"P+mQaB;iYYWQZ(`4(<^ @0dH<2q).(ZUh ك!{R!yiwuDL:Ecͫ\]=V1څTW$\ R(=lEhȹEs ~(dWmt&uvXKFSD9Z$UGmvCS+9[B_AI'``1 ;"04gZ$]c$RVȧ ̔5ܖb[m|m`RUNN $%֊J<2P\xbXi:U&L]D2Ux(ۤz;)jJf՚ c]luZZIVt0 D!rGɡ 2<㺉tR E&j*xH!+1-ME,f{f{ỳѵPim񺋪\v; _unkV|xg9ЮG?]S|$f4&gӳh\{h>?}K4ސ7z?V=vzQRC2L΢ yBUiubiFVx&Þ4'l '^Vvd?{WWc}Y9%RT<=żg;݋KU\Uv}˙$UnRyI JFtNbNykنjZZ& AKnX U%ublHI]n ۳^P0ݪlu[K=Q_u-JI^R9H)<^ GNkӨj`1|C%(1.V5EM6JP8BmK(!6E`wVTCIHeRfκ!m/ST/AQxV_uZ,vIMw4 ,| h0a" @O2!qի͙\a+Hz['8$(lEDe&|Ч<.^K;OyL+Ǒ31lt*IԠm-QlEG!b/PU~UϠXH()œZEU~) .09s:I֩$*_< V9W^ 9xQ!n( !fwC;_ % T*FZ=xżkjN(  fhA+v,^VG"Er"3/,wߞ+xWZX~xyO F&yQBz"ijt71j1+x?stz!_QVw>=yt~qyaj75=׋˿\nwT\\oB8[?Vd~{~{A̴H\?=h'>'!yׇ yfl>?RNG p WU@g#\u@fKvrpѫFPpnv1] +x,p=łj]PRm([,dO!L\gԏ#/E%YtIR2>x%O'SfI:']й ̕ir6< JϠ7NZ<ġovT1P(lVn4=V7ܐTaC>9;uU pNU!l@ӳsgV9D1(sZxL#?V45sln<H;9+Dы܋JPGagtcT[CM~Bn~}{L^bxVglN!,냣;'K|#Ω1CAnMU.K盲Kr\稦3B؍NWT sV8S :Gs}f7 jt D΢Vj/7+B;_k[b:?8V~WjR>naONOCG9:ws j.%3 F8Ŷ G㚪F% -yx5Ǎ0TtquS7V].s1ӷ_罣:˟lЂ0s߯4sj1Qm"zAXw>bMot.HK ն$rbc=!СHLEeێ:4ُŠ..Ce\EzU3ե&Fm4t­m/dۧ~6zl ~fw8lgfI[V5v(U!3\zvD_B"p&/Rj~̶t^#w{VL?.H53ya ю&vS ˈ@3gCԕw rbJ]AV1yGFC zVZVRbZFOIbQmvs+Z ~&D=;5`qH.Ds٥Q)YK)=eEƀϱɱ#g OW<$[AnʍdGދ2XL=)K*-"ܬ"LR$޹5|#J&|r/N\u*~ w"س0w[闲zn'`2/lgO Wg<Xb^cˣ>{+fbu.GKUT9h{Hgv)otQ叿 ND |qSXN \xܧߖ 8ӡ0JPA}6؊JQϹCA7juHյCi<)qO'*|a_UbRP HS)寫ؾ*)<:& %F C~l jOqn";^ %:K] vԽq"07D>nو.1"h5##{`雋,6=> T>Ib() ϗg_p6^}_Q:޻,g{.b吝\2LyW#]7]$aҭV..^cQvaQ~hW*)x>SvjO'8!l a3Nhgjh0hέ [}Ȁ1x5EB6Ʊcu&MiaCOfo#$J8XBuW^lˤ?W.i#6%mLY)3[HO*>JhY`6.(ZwDGu}3Yh{z D8 Qvجh:c:/*:l uިKaA1TۻɌeJƕbP$XWbF K[%Rr< [vp=JD 1(&'3;g?[=r_8v~52RdZLL> ye-Sks9QؠQ#p؊{,$ErɔrKŝ>9&>R-!3ch kA6@yˉ$,9vekC]Nw6)tcf( %J3xSZۭ"s }zk6D8*4\V HC,=YI1ɔF1l˸a*_E]JKŊ%b( ]%*(E@sŒJLXزO֡իN^EF ̍*Dt^|vI,uBiGࣝ]Kb~(e ˍ uN=x)FX0F"qǨee0`KJad*HmP\1NaWB!ڊq=t[Je%OV%!Z#F π)nO,̟Z}AGsσ_;()2-d|Z!%k o8R 9:VtPR6TI6P8tQd]M"\bJp8>gU7|*\/K|#_t رP? . r=S-v ̇sK/ wL,O /+%vc[l/;ُrt |vٓW4!?Z,Ԃ?=:|",q@J HdQ€H%wL+u R/13躷xbA`w.HQ O3&?z#a:mcK5wuq:6 uڇ?mDzDJ_5Q[Tl#nA{2 h@:s rnΠ%BAD\9^Cqv*a^D@<7{iy4H^gM@XPKn#k}kIE, 3!57Qz/ד#sa't<95?umZq<ؓx\\CȡnʞVtG@jQ(QGm.tyg8`̌, puKx4zē_u>rjύ4&,QUp=K 0>S!2|=KX5}%HXR[t^=ûŃ5`yN}@`ƸQ pZ# AWقev:l>RMMKoⱊWij{g~ŧ+> ?FuO#gk^ ,d]oF?'Ccѩ:B{W?zb"Lfoi_v6t1]at+V~¼? Cyw4`%qd|WSfR=հR]%Gܲ"W>Sݳ'/сqwV_n>gϗ֖ gx09/OO_#ٌd,Z1W nYEq  />gi7H+1ΉhGψKo+Ѱ.؋#JWb krbC7')उ zLGtD)e 2^vSpܟsgHŗ Q"4Z tWv_>b\E}30ڈs] ۬; &f1~Za g훲{S]z3ԙ(̾܋&Bܣ&.QK>.!ȳ8׫j\!*Qз;\56h2M0 4%јܢ5Yg4iT|̮U@)wl!s5L76?ZY@oOsf﹟1O49Tщ:(O A=(DK-㖺8ܡd$~q!OFFy /<ޱZh793'3^d/+yܫY8vͺt4KA$iV ^T֬ALOぐ}:Dl'3D.Qƻ0 d=(GD__U[~fϦlM$>tڌ}kmHzKbhP1T\!e,5n58yBE|Cܝ(GSWs1RR K$M>%b`yX5-c!#DlG`Pw0Z*DiM85jf\iKxM3o@ʿirnܾRJ[J.n/l9+$KnJ-yi%(R/N1catHifÊ^`i>#C`5icY$vԖxr~zxܓ?'ٸٸ /߯ŵKì*ɫW޽//w< D>`~m:}qY_l:|,obڔUm{şkƏz>Իvr~#YN?T֩QRbIeάWlJy3ڲO{4$_b?Rs짲a$sAT-Dx,#@_h~xpxfH;Nnsn//\b4$GzĈNS1JdzsQj P{ x Jr(N; [pŇ3ڒmT27:ldF*<#k9U|{v1 f8U1!;2S#WY$sI:b RJ0 sKЋT Q5b Ő4/-::iݗD T(d"J-;0؆]Jm ?VFS*Qsގ65>Հc#7U#6zVCCH VMINFkkJMC˭$ y,>tR[IxyW?O6VH\JJTz)6&1OOр{iPSS,:毞S  -0knpϷ3Ͱ0AJrhܬ5@R _!D{%uIⲘ`>6"Ba#㍃p3QFэ5΃ĤΟY93 'Iw֧!O'%hl4պjG@ȳ#Wu*h/s)E ~`)|Z+4& q׷ ^Ȑ$-k1FAg]=U>^"!J0Aޒ&<L #~⚶ Qp/&r♂A =N쌱c6'=6I.U\lF˗aj^ȄU5e`T5*X |ٲ\b-l,زдCŠ]ΐdT }lio=lQ)}oMcB{Cߏ\fX?Ͷa[5. }Re81VfbG]#4I|#CL=N;$CQp{%]=zyDƼ`_<'yB [A.<}`|Sˣ+X]=Xe^;~5J #ѣd<7'}*^3cSD}Smjd߹Afiyxs}y2z+me^~r=-:ǎpKZ5!VhywʪbUFU mUA*9ArʞB.k$d|&¼KYN1[:ϧYªF_5O Ǣt.E1۲S`TkQ3h-Uu_20"!s4hVogzʝ6/#VMCv\k af;`I**4H-8[ m1P!2dMBe-* |Fa es\Vc[\x7Ë_ SDBC?>:46ba@xLb@jӨ![3Wdndzy"TJr2h6,LJ?/ן.}dRwq˟YS*lZS$ёW;%f"Q[2ex4 5{L(Ĕb gYOS{wF?x\t TI2j83q7?^6u/3;}O[1[SqXV Μ`{.c{wޱ3?G5P'%ɢ.$ :qvXۉjrZ(-\[VeGieA+%G6y@zɒͷ=c xLٖaVCT)&=U *ZGV8Ȫ$29 *&LKPK8ơZ* A9Fg1x|zuwWMٱCWE9r+Y @e`hwmjmE|S;z0^zJȺ y4MUkcd/SHg>SjU,Xep.R*3RQ`3,@YƋY}eفU oC.Bψ";/[-r:9({ƖsKY.>r' \1ܱcxP~BxBqY O.C(V [%~BN7T+K+|Yʨ39KibPf"bmpdȉg" +MƏ*Dtc&Jwp rfbʂ̫eB7Bx AEec*  +-4܆S/^"zL[8OACc"G73 k,̐Π0F!'I?WVmQ-zpJ<(Ho5ɏW=ܙqމh.$NJ 'dgUL!`4t/4S2KL-N`w֡x}w?H<|;ϴdwfo MhPG~*+vWdސtJ9qjZ|s69Y{VAov&<8ڤCvdpVNFHSK^.MW+>U"YAS]N7>uij*$i+-Qty~4fiAlڬn<ʍYoDB Yen3z'j;p݀{'Ms@yû$yʹrHMs27OfwSt }Όt܊Yţs-ĢEM烶rk9p/- 9Lp&hfL?~/]Bq //i(M,Kz}Kgܚp.[,YwoP׍Sl2-G[%p%=mȻN6\ W5~;q)5|gDgB,A+,oӢ-g/ ¯595V!z?BQ{ZY 3vlzcDgjdjח`|;n;}sߌfw:N #yrv:ph7:#{ll1\k1 ]gC9<(]^͝'ii=' Ft1638U#4[^NʁlڌUhז&f3ODwU++vH$ݘ1D`яDG]G/83/mDY*zݜq:֮a$^6bzvbŨseI@b]Fo>=1c!vztۉ9 ӖYRuj7o:{Ԏ+o/ځ]E<PQvm>o7w[&&Z%ʈ&oŘ&6G>Ȳ?ׯ^iun;5Z 1VLrk_b.SY;-؜]zqY|3~mWU>Jw?nL>*He5mЌ.MM}L:v@)UC(Hd[Fע<[^p@g xsKOCâ.yw9jޯT̚LAB5 %fΪޥ:[ U{YR\AQ25Ft()2du\+s!.G@5|F ^b5v.\]Og^:.UGC;<{}2NUkX5! 1<@)i+(3i M#E9#|NbŠW ͗w׿j3/V$[{ 9~J5_jFo!N 7> 9 m͝8:]vIb Oc}2?~No8mEЗ.:2yh ]&.pWb&[[#kd)"`ı^9s1u[Lmkv4q쮸Q%թ`þX YjiAa vRs57CxMݽ~'}\|UlHw:q(XM%s//nWЭcaZ n{~8>j@uT!91SRH(do{~mxv t#[$cCpt]CYx7"Maw,7Qˬ+SUUBldo]Tv?}B% WJU8"R't>|h SO0L~e`lg6˖.b5Ćөkv-MkC%&Pr[ %$cEKxKpU x;AyM>>~}8$y%L# G'+qD^]#ڀku~2Ձ/ybPX-qR‹C>ƶIŽgĀD*XZJnX#[fܨm6^ߊbZ1gr,D:4'"$t{muޖql]V|vW,` >;cT}i&|v@Hj?zh9]Ju&Q5 8P@ot&>Q\S[vx%O@+eY$r$+YN@*CmS5yF~Ȏd֒'bW :ckIxBgVZ2A>Q19)[lxZ9{r=$Pz9HR Fuj|ۊv4UߪNonF;͍;J,T\vp`b⩅V׀W7M {w=71B8HN&f}-订Lbm?Q&hUGpUT%!H'x+Y8vO,^'J;,4=t,H*$%wfDvV̒%Frk6 cdLfFYysF"nrkDJZn\ F!?[$3= ([-D#v-fVT~OH\ytf|nwyi$Pʒ?p:?bMc)´Lfd{@˒{c{$?ńN)}'؇Z5.o~i޳ҏYӓJO%ޓ'_x$~.Dj]B+kWq{U~8j/|[v5#'~?k$Zg}2kr^2=)~gڧT2jB~QGqgCĨmAr7~Ŝow^ !6/@Rޱ`/)seU="DDA$K:"tKU5JJiB0%c7Jb)gn?rA!옔OE&.Ol/є!;Xjv%U4R  NhKЮY%h+hlTLC³( T'hr CsZ:0_KPi('"-j1C $jՑAQA%2dT#\ȕr1ܕ"BWy̗D}7dB8A(+gG (;{9@ۋH~|/5Ù}yֿwnةuē7Oeų>KGiSNims=#OUM*.` -E"&rZ? dČ{%\B|N(n_흗2B$:s[.]C%(mcu ;NIT-CeW6NXz|Vlo?6~yu^ WIf% V;BF@/4`wܹZvVG7bp)|UdS;bDnGfӿ !ɬ+0WM(N" M0KT1 4z Kd.x@qfVI{sI{#j*BJI{mohԖؘ,8 V#\Uwcn$[´S|ֈWbp BʔI }XZJ1(S{x^$a7"$toDyJzdʧ|].H!({d.[G*ACT2 Y$I5Vmܠ8͗k.3E8AZC%XDAvh`3 Hi~UqF&Wʷ#eX05<-6%'?> N RޭhISo:5^k&S\iHL[`f(xkn_㵤.ϳ N|fPfY VfKƘ}V=X3Kr"oFEml^^fINl+rx96+JY9A9$@Z*8b\i)Yu{\B9ߘ\.UA=ET.{uYcUdlwl3e(Hij}&d{BpʖG%ЪGVx&-)|[wU,E-Jb:e@۽ѸmsP <{ATg/P61!td 'ygoq@^a;qIPtoH}F..㛐CІMl+$S}$E(.MCrtR Ika ۉRO5W`j)*)[bIؘ;(c 餲^kJ$rvPk?Ty| 1A ,5*FƯuh@C_T|2Ɏ>=1m˯\qWY QFw?J"x*iȀtޠ˨@y#曦\yVm/|j$U)h*/&L;(#wdOV8ڃx&`kؾ67u75* n606[Ќgq8$zaqݽoGLYv#IMn 34hJJ%xTA@Ft}~)S_= ;~)'Bj g畟U}3`7\n&?/_PE-r8!x#:kBO Z;ZP>]A\>eU}]yML0ZODFߑ*U t`ttC3Ɠ&ɩlw>.˰8ïWvjj4CH 6*i;F$9BbũΎ`CC?JH4tGVT=Ev\"j0J!a <4AI:CErayT Qr > kDTP&pVL})2S&=|gs%G5*yK4 ew2_ӱ4ݹ#oѬC>)8OKE ݋6Z(ji#Ekzpю49A6ŶCV%í0ʌQ^ &ZA(54;n/b!dI :"5o{_\F1;i L'ފ IT¼ 1$5:pki/Si%4~Y`ᦍQi0' ZF)U݂W;(r#v:_ɚ htGG=\~Ǔ!.&60Lhe2=rO-w?ڈM!^JI4XhgXCdD{t]yXㆄj(xp逽HA?ȱAxk'D5~(!s!_q8bt Rz3/1oG"6_O\KoU;o\._~eMR+0)Ҭw rهb#QI?n{%MZV $[=hPQsDRo|[%uJ"i*/]dܘ#Q{[('BGORBS+N0ӌC>N輸 U%$;W ȡ +oE9anp9vR!t*5+P\T@;v!312BJ^T,H(HT%㾈,%ԫ BY6N ϳGJ̗US-'\,]x*} MuJw2a3=ybs ~n^դQC_m1k?1[\쳉mC5Yp%AY}w~3.> 7L)š9ًT[b-H ŏ(,A8O`9q&ǒ{RGN B:iRjHؕ!'JFX\!yc7R ^ʱD D#t4$1FPnގlwH6G;_( ** Ln?_0[@z}ɥa%}@TbJN(U `wY&NJ( (wf/[K;+|{\'P^A22ˀ[oFy>ol7m4(mGLJ Ҡ":蔔brB@ m^{σUjB˅!+JD#zR$2xF:fK`P6P4FF"p e ut(/!~n%uHd ~e{LZcBzǧjrJKܴj?- o)s9 y%I6Y|0EĥC'#p-k6@\Wj0HS!|%%FlQ6g8HbJbn%ATx425AAQi!uڅx\n\dQcSJACPIk"HNh]UQ(C"Jώ'S/ O&G ' ΘFP; ?Vh2DʵNJYn*<}0TyIVmSIv=L|U&O}*n|?UNRdU\*Ա#ppn =O΂t{ RV..[)ib::lFiJ)ʸ`q͢(r(oyFrlaK{V%BYCȈG0$Td(teQi.:*&e:@n`rS,| ;C ҠB>` kHA{6^>XKG,@ O>_͝V9֌ŕNr[8'/I,uyQ43T ~}F('2͉z53vC[c#8ϛ{FD,0$d>]__ Kٟ!?\/ UUy涤ݙۣwBq#B?vxm4T1%/®O&eiU&y0aرs&f]6]75!\ޜ+Vea%kǛ){IWx=u|@waAMYWH8ۛf_+ƿAlˊ,Ur5aK޾g?]MWR0盤uGSZF#!RJx#C Σ%z2ɻ+S۹\0"+hRLWBl:o@;voNV! qtx{K4Rx(OgƑ ykr4ʡs9MK1lLҀЭ<ꀝTigyytQG"h^+4G;c•W ,%O:yb0I$sɧ"P5kP4k_|maE6cUfqta/.RvLe./ڈB>8(0&YXR>*tSA!'An Ww\x3fF?H-cUVf>t:wo_P͝u,;p';,,pŝ 8J??eo<G>}pz|i]{zSN? 2 j?DeB">E@]">@S=Z/t8pO'{jQaڅv c3M7,5|يm鵱8Y8y9˰c:#DV!k SY;;`+_фjFjBT, ux+3 u'+َoK.g9hPB]@K#Z{+#ӏ\\A'>s0TщX|(N(ޏcuPA8իYchP]=BeeQNY*eх ,VB91(eĠqݓx߸a("AR[ɦOJi7{q1&_/}lmň{½MA7ₒ|99p](\ߏ62hh?3w4ܼ.4`T{.Iq<T{I)B#$_Qe2mi8{s[gh{ IoaGo:eV'>@?LXύ,îbum񁤧P- [ڜRs agյgU-񁐧vhXnDCxS&K"Ί50>kPJjz+?i|O'wn U>gIT& `g/_ƿ<:xf泻i%ZK^} birkSBS(ЅJ豂C 95bW Ool5@l-%|XR/ׂ@gZ6_q!MjU|r*uȁ[FsYLc{7Hhe{-YAl,<==Ƃ;ӁIgHo֒ZrS4c#, K=oit81N,=B!ƱX%VKD`"eaI*b8#QFk(Ӎ%P)V =wT㛀-s+ 15#rpϷQ@. G5Qxr΁\L͵FXI)L*qW$)n0T5ٔ0hv!:H~Z۩Vv<~_kW[OQyL]hmڜ`20o=l=6}c$w;~je}OYcMbzR6+'с m)v }etݢx-kcGZ/V 6G2)P` Ws,ʀϤ s"CmO/CT ~Ǘ^<];' 1>Je><6^6z9?Ͼ_Ou f'sbRsX_.o*ח>ǝ٧ϛǺrϺoGˍ?u|zv"1P9axqO`zcw-vRxuf *=;noקKooF?7rqy Εg};dt&Ô?n4֌.a8Me9<*br .Fۛ=fW_30N~n?{pμ^NK"6Š׵:zO ؞$+t|Wt(>?r0^o>>v^_ptdD_4fK?*Ξ5ӱ5rʞfL4ٷ( ዼaz0*^rjp9O`~)YQBb8~4^\elp5~TƋ1[Ew=`s9/tc`k-qĄ][YWj10/i,*T,Q\ᄥ%:>J$ZDZ<;ֲ3*DoKRճˉT:ɇT6&UA3 u(*cNRdI${E1%q술hH pR@bR\ ð6^iƔgFf}zPKҔb ?rH?**a1_`(ܕI9JLN>X^XJ)/bNg#"2\q9, R٠ C<ڑIx~h)?a2Un5 $ډ)˔J%4'0aBxH*J'K%[+z)pPz vؓ@L"̛S 3e^4Yp’((C,}eXEvbZb7;-qQ.Z͑QH{i5V7BjZD0y†,qhJ#lfzꂔ8 `~SCZH{FB4B_LG@IO1I MUG{0먞BT;DR3B1 ᦵzڊǁs VUo*Ջ hY*R2&K+--^ &Nh3oBCr~B2-܀5郒Xàa%h;S=X06B d{?jV=05kY)y<}BEJݺ*xw57Iwm\dgΕKLY2TwL uU0N-гF +?I>:MYu}4%T:OBUcItE#[4+pn} ^~dX3[;t4ѥuAHY?QH]ﳼVVFO4B5^7U*o? >/JɚwTyFMۿo5+7Kx7O;OfR(k盱w p' Do*}ۺdoFOWpX-竡`""[qqvO~W[/z7W~~}M\'~5s{i1DpEؓvaJnf'9*(f\WFp}&{֩8AGNUKvL}Vg7N ǂm/ֿ%3(\6/gi@l>GwzI/^܍ݫ~<\tN<]DBB41ф&2q5+1rK=ʕ,j(ô| sE9j$ 1* *EXk+iHZ 1J0-_!-Nf8$ Ԏ0S p?}*JrB* d>,A u7wXÚuN}Ќ~o%v.),0y4EzzOQÛ ~ѥ0?7Hcm< 0Mێ&ZܦկՕxz=yl7{'lJp$M廫\&1'RRZ[u;x]nNTja~\uCv,4d۴^i]`x6bQy,Ҙ 15ԃ kM iS۟Y'B~`ERZO>@82i O jPjyٸ6x;జ'L/r5 ۷Z!\8SՍ2|3 r8֘zg'˧??pc$p!ԖZ$A4 cY44[Q3j{1D&4rdj."|%Zs(8/hUozL D* Dg#oyP'> 8xr2sXw[A/(dLf.k&7>xa>g/N9*=2-ԕ/v~GW&:1EK֍;w ֭)UT;X S[s֊mWF:%;֍ݚbPIuu;^2p֭iG+[h?\Eoe-^a߶~ o8l^4]4ZY0vd5fw1.޻bzo3PA³ePcHh+ɘRKY% 4e8̈jCԼpL}䅻2ߥ\hNkjJbװ5U{_Vk+2qF\-Q9&Fصb'RbFp@b !E1~Z+08'S |)۬WZF_ MK5pHl" 9`>ymͰ5 gclzc+y$ny 릗W_F]iC9TuPp蛯eւptq{nWnsȨ4Q4Meʥ2٩%jD FYF%178xJ0FfC]ޠRr/|M= nUuzkЉqI?Y0-7Wj10/id g*P E ,1h`!jS [/"̌: 0E;#z_h 0$ V> * JmM~cU:ZVZ ?{WǑ V+CX/0}5y Ua$`FVjvfe]4- ٬+ Xj5\ۀ8̃xr5TP[PS{@"wBfftBV qJA aآ>Ȱ]c~dX %Z3)OLeJ5e^GhpScL8Eu18Y=XI*s&fk`HPg yZ dh@>>(,\a-b EԊb}W?QlN.WAJQ9ypz@ʷypchWd+Îw»b:cxW VޭT6q-)$_ER RTtGvޭT6qSgQbqlʍɳp*y%26c28]M@+kVE*$|\k>wL@7K#YeTM@!L-F &9KNƳ Ht)gpgsKw ҫN< -ȚK, Uҵऌt4&"jPD (,.g;zjAj" epzQzj^цIS|T":M(CX!wXSśNYOz|aAsp2vN!Kr-)ScQxf:+S)2%jR$9akslWd8cUBt $vنkpvlxCE -v)um~l|m;E p%?Yz񵿂5O܋7Trm> >wѪ:Lt"Ȩb,B'LY<=r)x x$Z5g,`iWS$vۑ,Hg~oKr#PwR#_҆smjw) ^cZEUp'T?T; !wIE7i nlN=w7q݊"VƩsAlܶĈOs`zLT3 t#ؑPt%,BW֦byOCSD,tBҭm$U^D4x^dWKF+U18wmߌ!dv+ZO~[fJ _z|ez״\"Z?͉`h{.ӏI>EUSTOmkc8 YGrR۠1vZ12m ƩtSker?ҁ~$zĎ)&}~qpAWH+Y:H}$p98a@#dp51OBx2VP1&$/EA|\Ww_H\0\1%j$i+(*x]1&YW4xLlFH*fU×ԡs< A[-+'`WԲaiFR#)y%ٜO! ]kݻe@'27u\(x~%:l+a@Ϻ|xr=g ǰ>lsq û_?ۀd]7=\`p`\]^_໌aY(+L2M'-w5%Jgo_sF86$.m9!,z}MRttӾo"l?, 3hv܊8I`c͐ >I0I\!Tb2_<$MTْObT9}{>gb?I̓R_a8ICdOZdTr)oKf 4[m-#NxTZ; #[{"vY `3zGsju4 ( "{yM9aglPInP!o\EKt yբuw+ EuJn 8݊MnchW"b+Ύw»b:cxw SfnŌz14䍫p Qo;}UZ8M@S]/@hw BY4 N"İB#wJyPzƼ05KR .k]' hr!:6J%\e2 Qp*Wp\Rc Q/Lj_pǺ{zv)1VkDomW:cUS(t>(͗;[:'O ěg.'08I3+ΓY^Ӈ6P?!S^{)T;$=6Cyby$^[I௪k bH v_," # N]?(q$c vJ[:'af5< 8*Z"?X͡W"I ^b8w=: KG*R\ኁT<q*DL)ES"vW zFJsJXҴ?S+p5]\I.^88Xݳ* 9´K"17 Z+a C}M=ιW?`&'b dFc#rW%ND*A@h,8k~A>gH /.J)( 'pbrL1>NLݝੀ<l{JDո !i,jdOGD !ByVȆZ#ƔZ˘51I'aN v87jk΁E)-njXckYy ;Bf,l ֌[hg>zYa핦(xx#>1Zb0^9 AwcH?1??1mӦ\+pȌL1 f 6OWCГm}_M bdpл0Vuv$$)@!4 `ߺL|y= zZ (_@ v; y7*2>}` *_`o3Ѭin?7|ipUˀyepw7:esj3>rn5Ԫ@fu?7[XG4Nl'\T،kK 0ͤX.<DB>_;:~AVbbJq1 9^dĤ̳){j2 tb/Ϣ܄VZFfQ7Llfh&.V0&'KBiG >Uer^*kYZ%=K%%i=s<$m5'X]vr̳5Xi%U{!xp# =5pCz1??w~Q,hp+w힞]dT+OMz5~^҃#+UXU ]h')ٗ, g]!k gp54u~\4w[]ZQGo>>h~a^Y?9g{+2AKZV>UC'4CK+ErܬNFqܬ,01(7 ( siYŰLL`֐`9R8[F svZyFja`ҁ%VJ*UTSNQmN$)#r2zN3Ij}4䍫hNIx3璏TWz>‰#4^CM^L$-2)T\OժR&OQ[CȤx)V̙¡V4kNƄȳClkIaa0Kä VGKe$L S`IbtBoDœIS$(6*"B7|zbe&^*3Se&Y"l  K,%GEb}h4f=0zoAC{uA*w @ \< Rci//MX cdE"i ,-:H0VS_1)Lӕ#wنϗwKDZ,UyiܪL(,麆0#QVM-ej/e)c:/ʤNt^u繴hv2xoےHf/UqsӸ]l,Qu]pn(:e. x(r>߾͏I?.$ޟR‚xSqSo.77cP 93gSNKءrnvKsiPsWDfjsVx!U9Ⱥ-sHK8Evg5V:UTwW#dp7u__+a첪 8Ŵٻ8eW&r/qȵ/ A\aBQ[h!7gi"Uw4`r.՜fEgҜwI?}evRt΁!V:> Y9_G\,OS>w[YkS Uګ)C9BdgJOqZ3a`\įr%sasCL?"b]$Ԥ٬ϳ]AG]`5Iyv0alO!ųIlC8w|v>yZ~AAWheKKbJ\rIYtxv8q1S3:jz9H;gcpj]ns69?7@rmR2١)kXeۥYA*Y<{RvBZOGSBJqPNi:"x_ _,*~\v! x4 9rt$(fL4O22FO!-!CEx2ipHhT'K5+ w: a-quJ6a$ i+l uȬ8RvcOEIu2HBMlUxh 50YuE_{yBOp8Bk ΰ:ƻGo1S[\vAR31ڰ|?5RÏ;Uۻ.؋пL ",Iӛ&{CI)IeoX|́h{b ?EϪ^泫)D`O1uT ,Z9`poSĺ91;R,u:1PA%I@'U3uF("j: o6ӥ 頬z}i:g3'Z2wϼl~?~ݛo8YIjf/N\vJOjGk~ߓ~:3=9" b5JW D 5£A3J;X8+ܕP%nrF0T _Mʿcs tQY5GABu~2L4ngStbVF*۲@鉿]:;/~w/=%ixC޼{Mw?.BDLYq@:l%dooo3AS̹Y|=|LՊn//8СH,ӓLbs,Ҥ;>xB1i.OQU闷,>6(Nc^x3K'o{0"@<;+fOa뛏6fj2$($)W(º鯘Xav4@pAOVrʊPL5- < jd+q^z8!H ar'Ze:k=ԂhGd|IےS.;sź"CL` )&{iB2HO|{ Cj(&c) R V. XC:DX-F =1ysisŤ=R$AN#1XU;w.1I{͠ >eĘit@[4]C p3"EkX,GOt#&RPQ= "Xth4~4Usȑa OjDt>)R2&YFxJ7(ijՅ080l4Vs f )AP"miDRpBV/lPʔ]D6 X H Z:u`1.eSe5"RN{LuCAڃ@i‘// G3DB)r D- '{khm2BTJn6nkӜ&jAH*֎ݤ`DL^Q1SE$ "twG ) G z",3zE\,[o1H>+D<__rђJN|/,|2Zq61}e״8|{y9vuP| O@K"DhY"Q=Jj${ߜ"#S$zxPn>׻8nNբb/ALJ J#Z] ɐX4O50^])2I(سx9,^=cϪc-RYwV )C>HJ B +B|B+U?w-%'=^;={wEZeϐݝIXK{9 ? Do~<-9}sאAI퐹Cf"(6L"LFKĊRtZUJ!M0k8x.ױ:&Xcb{c0狏|t23s>YC'q` JQ[dMO,*m`b %E(G’*'fL߅sbQod鼥4se4}2\R6e%7)N'>[|e@n,#݊$m׳V56#3ࡰLňhc Y8㹰$H1|' @E-ut/5.g*^Nr-uԮev.Dlv&jt^dwj!((G+IH+!CWwBsLSͦeĒӤd3xvMb㓻ѳQ]?/vr+0V!l)%ca.QZ/Vđ?=_m09h}&>8]ݞ&]9<~VƷ.&?c9HI| 7ͼY0|\Bz_ Ƶ4ylz~2)* {)=-aQrAӒp͑)Gmh7{6˃v(fݲ MnmH "H>nWO ry":cnÅw&4U!!/\D3dPef55a n6+Z98DuS1]$xt)%H#d eal~Fح[x2VLRI(QάDҩ_Ùz5I`ǘ) ^+6(!.wy癵SFSAː /t@AAWz źQ*D+k4?;mCB^fɔ`MQ EtrǨҕ[6ڭ y"%S nQh\NuۀQ"wݲ MnmH ,y Yb\r=q*"k n1*DʷXk% A* /EiYP8,+L)Bsn coZ R"f7SM3:{7[ûA7F=~U*{s/&{E4K%n >˃v; ݵv&4U!!/\DdRhv>cnѫOf|f_~ &|;xζTQoP$>cħ Fe*j?F|Z!^o7$>cħ!xG)(@\j↱[4#s弾Yo.̥y0/=['ш]y ?=:p$t䫷?>\^x"a$J I;S@RMiI$lC*e$,;1@*{9 ; Tku_%IIewc(HZ]% IDV!nH̅_v@=l1wa!|"aVC6 'xҦ=ٻq,a;ٱl܇v<[N\9z>S.mmdI#T})XSDxT8*`擊W1t3JEKYT$BH,ҩ~͊zEIFbpZf\TC|)Jr@23b$Ցw`)y}X mHJScSNe+%HE3pi< 1zn c_s: g|ԋ~\ }Xƭ *Mwm#v'?&)fr>=oe.YM@Kx… zWN_@2ԶG@X-o09jٞdhawZv4;{JW$`)/Cyw`jT8w \Rs~#+EhŜ=КRP+8++?}.lRHΖ @KFwОz̶#5(3~nc]Pp.Uy4Z{W\V ILQS* 'NDq͘J\ǜ洺 U:YUT)@`up4.d-JbYe@|BPC<TPƀ)݂i}oq )|0hIm_zpƺMYev\ >.*/NVPBjpݴ{ɚOS2)h4OR[vcU}Bs$d;c'&B!q@U!B{Ҩx\jHԁ8oc*\nZhUv= j/ ,46zhNN>q.P8'KwU3iN4Ft:&\hcCTVTV^ 9@s俞$> p;Jec֒,`=^Xgo;R򍧮 ;^fffNZLp{_FCN+Ԟ\HUQ{K x> $t[o*\\4$@2”TfQ`wk&ջz/*MD^ [#q >6 n#Eз"H :*>ݜP| 0 $-4% g:܊z\ .c :<0nd8|4UB _saE4>tMq; ڰ@ _GOn̗_"|]s}` WtEnK.Bx&$sFu*c 1 f:'BV[¿w۝ޘ>N~;:]}vI6V[Zv7iV&|ZLБUn0>㣷.h|D Uus3n}G{k\k_>/vme4]b+_Mׅ|PL} G09|J lc5SWSI3=cZםi @o]8z$8z V᳷ Wf!SL;` {60a g3hvz}D NRblNe1vz4gMN@R}_ w`˚ tG*)F$6$M?VIwc֒q 5Z[Yis ݖ}m hPݘn;X|VK#\U|3NUBԏiX7:JW +- aIY|@-^[K0)/Zr㰗zz[쮗 sI^_# z1L?/&f-Eh!qZ_/2qrG B,{*b جvFl#APZ_ۃ]FSDD~'w=B}B%B^M10>0h6iNۺf<[Yc;hkɐbi"BQJ ELs!R`aY$crdD*4) ѷVC| ":F 1;I|/ (ctG8SyR,AщOw{ɻ[;բ< 4?&}1ȇ&.:>t?u{t -ŷ4Y ~Re^. h\.K$V 9u4:H'i9V$ -MRP$HH*E(eO[Mmi{ҟ>ܸw@ǚn5ZQ3jwX uw4a M,wW-t9p0 G'ZOgҵeY9[O@NEƐXnҦ26elt1SFtpkԈ nor:zwyg|wӅ2L:{o^v˳eY1mg(7˷o߼}bvgx0ѽwulj跠DuYɈnὗ;ZՍ!>Μn7  ?'YFwn:LfojO_iѹq839dC|v= p4eOz+h*3ecw.c Tn;Tg5(D!](dmѓ__U =q͒e8ݟs@ȧ\x+"_}_nEKWBOʐ/5,8aZ풮sb#Nmm%Ҥ#Ghex8e…NM$R[ee1K4B | XXg ՖQORbB2X)XTc*jI"S+$̒V]IrZ,_;aB= N~Pw_iBpvًӎ=hǥҷ 4+XK"W,~=;09YP BP [\g{WrkNd̐~>]v5̭H(Vx7o!Ag|$p0 CUic) ˌa(IJ+ A)j|rL`PVp]g4Y htS8)4N1gn B8v c$V(oX:vPLB"BĂ^X>to,UټjYq@-%,1W[esǭYseO[>dJkʯۻuH%h[|<~?^X2^ ]~<&03ps8:׳o86~n',%3*lK6}`ҚQ /&wO{Xïr|P؟G%YPMYp{4Ͷ->[lyʃJmY3XJyyM+2DSE@c RcE`GZso{-!?0Έ'5},9YQjG$c[cgprs!2SwX7U$Sݺye\p6> fO{j3K3{bgu q`gƭxgm_ ڿ'Fz록^`Zp9ӒҽT%7لűT$'Z ZkMlS4%#D|qցߩz U.6SfԼP GT]qBb2[hy4613m`rcLU4w)OA$]<~`T9/ɬOm74~tz8ߥ+E&#V $0PR>)LW&>vAI1܄`v)z3^Z棦L G]f*5ҵ 75LE齧mom t"A^棖Qi5Iǒ8 XoSf6pDpP;c9{]3RQ.67n#鿢[Kw]7W5֥nJW.g,E_,z1)q2e[Dv7Oh L mO*#LGTGζ;=}kxznJA0EmC_}&7Og~_򿏗g(1obmZji1yΌx8;;/*ь~7P f(Kߩ U4FDsuT}BtAՉ}Ft{B7Ԩ[4Mѭ U4Fgmʠ> =acZ7ݢmn]hWQ:S8*v -VMT'm r@5'ChF[hN-CX;Ɂ]tZn?fˇx8U %(3Fty4̫4KiҬ1~yk_n7qc^1ta1]׾B 4VnF\бM}k_cmѴr,A\>b|0[?(?8{zT[?Nޔ-\W,l}gm$X^| TY!%ݎRʓ:I?˟S;+}9 AƧ]ů}=J/SA/v`ة<↞Xn.Z_b=OX/wY[тm$bu!h3wP5upQ>*x:8EnXR5 dd5 Azg@Ӈ7>|$Bp2ܼ#8ׯw::"DC}8`㈦~9v/gl1I/dݓP޳5|@W}?*z<.K%>,n[na8XH@MժMx<@imjԞ@3p.nNX]7w"d7"+ֽ8)z=MWV59 s;ށ?/7:z#gԷu8TL`lfI\&11E,*,3]$?#<ɉ c:ZNÿ9Lf2{KMDXns"2a _vl()ׄ&F4Kx"g ^q`u34%%Ié4_'eƔ|(9LMur.AD<(Ŀի^=N^ӌ KŪcm?6Y׹K$U ʬg8r08+.*n4|وQ=9Z_bdge"7_켋NU;vyNUbSթz)i1捺PP /EQy o "LZspr@TRP~[mAxiZiĥYh*\j7"a) K\$UIq!$(!!Ca& Se9T1u8U6.&^PN$2I%y9Z& !/@Ɓ(>r@vR0賆ʃH0 ?}|3 63e/s\3}[JPZO&ݲ̀Rk+,N|lL>r ˕5TQEG5tS2Xr1E3ZJұ ub$F k4"H.E*ʦ3ƫx9]BTKKP={krG:Eb/O`i<}E+cVf}xz!8Vu| oHCOght3r -t>R [!^z{gʦ,ao#]a Efh{%(˳=ȐQ{Y;զxY SBQxp\ *UO].4 Ki'`Rr0D4ɲLՒf8"H=5!aaK6S)()-[>(d>5 g@hDw,aaaaudk1$,!LK3cKʽ䄠5Q<֢Qp[^ :CHt9#BK;#m4pjAOp뜔#Ņ0В >,;8#RS]qŠ$ =W*!'%"ZbxT!TF-MWr(1HqQoD@C)Fk.E_ +r(hgTVkJm4'\B!!i kTNސ>PBl;üՇg*ִ+uzhMԞ*QDCĈ'+3Ng)Sw9pr

~c4g0ʾHA\ޥF)T;| |#ɟtϚs(pmzP%'u\`J1Nv3.a ="X5$XPՖwcuv\cZ0!ѭOjWJ0KVdJ_;t\ߎgHW[F0ie=8B5GCDb49klo΀NR4ibM}fưdBz# ڝjH&R#?@NA\ܜP&H/ؓgUd/}fVՕ{)x9%>7ʇ W^mĞ{RB)LAt+ʩKЗpa됅M@JY6kS,:9{*2:}TZ(KaL9wZ;mG<2"LTF䞭Xu:s ./Ξ93+ZU*9U`XN\ҠUw04}Ӷ1[IJL?Vz'hR4R&01 NlD^3ͱ6܂CɈ1f8i\P 't(la VܻD[PaïfYhO%_r؝-("N. ݓ6Գًӫ|j9, "RserWd|J#4Yi""tC)Ck)y%b ~t/izwSv:mDU'Oʱ#fntK}oT |G%:=Q]URBF9#EYXpj8q1l2|b<>ðق!*ڻ.نbiTwH TAOm^M@>s[եn#J sJ?+Sаe#¤6+-DD˰J8 =aq7ÛmǛ7rϏvRIz,,\bjaCN>uqԑKyNoI^W(o83 fl]hqo';NRood D+y6j}ށ۩3Mav BPu Yx}ݷS]<'tv|$A;ʑ˶vOe0\ c\{c=Iv+KJrR\AKXO<KNiH7'8'b"Ǯo!]S@{58J H~l0+Ѵ_P:E+m ]նbC(ε/#pg щj hxD'K$طJ=DrwkLAihęT[ ͍0* xULlL3\_HpݦVrNj߸6@(!D@/ Gˮ}990mqvW٭`k/.1$8MM_:N'=g §Y1 B9B]؞6R&|sPY>#ns2gh>U,SP@a"-kq0Q#. 'zj!y=K c")t{{ޑmy{@'HDž6Ww_"Og;i]e r#. 鬈v})ñ ޑݙAK521 -2hRɢNg cC,:p)Z[!H`H&P_}[dEչە׏'# CBNuГ=PM]nf&uMֿ)sΦ_Fɸ(#M;띈|bT]ϮL(v)yLߋ2yvj 6aIO?]j (M`,dP`@c2B)?ǣUվ7KE4yuqqq,(fRb4\e$Ui DYÄO:~W>,Gzqbd> 8~KJIߗ$ԝa/'ՠLd "-EwGCZ6_egB+}Ⱥ#3@ @;Ab_2V]EǩVH%r_Gd-;{Cj*=F +݆߁==Œp;:pomnrLrjxxm!jf:@۫!@6i - Q]=4|ү홑  gWRTexMGFMxj]Kpx xP7L#G{܈=1!Is$~޵+|/zh!1NМpVo>;gA%su< pbj{Tw)98$NB9)Z͏j+<~L0]V#csH 﫝Z>x,kwdwtƅ. Iӽ% #vYx$ Π#j { 9p($W"I )W-[2@B;X t3.R34zaY>?BF#0}-x !Z j h&FO{?sJL J˔1V2b IŹ#LOV\Mtjbhu5R k>.rlz{ ♁tߡҩ'QkNG%_*4闺60%;RFpR+Ix+[Z~^ K͢y2,<8N QGm?+= x`2V[ַ=cdӆ# W+c^WAhz "HA#pT%>Qf)Q06^]i~;; .6j[+"(NS5o!iZTxku58-Q|gw?sa!G-s'b-rkBsſsM눆ыq6AŒw&k9K]\]qEMٝ@JM-$5p7)S䰖ۍz^6NtC@5nt!Sa)pr.$M 衢İI=}h tçfj'q LE˙s> URy TmGYi3L6uL>[+׬U7ȁz;o_|S$JlwzOU Ui2U/H$y,cJ2S sB@L)29))ғW$ }nO JsVm~@RFjÇ_{M1 d] I_@WS*χp9N'ļe#UX2"dB>0R$Aa痐ǁzƱgt Pjt]"7;Yɞ}0Zke9zo2Y%$X}#WNٍIgcgfvb"<H BXv)%R,GDyJʱqTk#^vˢI{ւrN -IPpDCڮŞ WbHn=3X0LBil =LPN`'nV)G-!~>;0IX-.:ҋ>p8+d6/0ToJ]k4`z3Kmbm?߼>x Ծh: ub.׋kY+~qjPKbZBd?0=HdHh}{LDu{u.ҭF9>׸k}j[,<ڃ6yO_uԲUmTCiim㨊2Jo#@ IxC^d-LUBNJHNiTt%;-o[Gq*qJx94.ՐYz{nB$)MͳXzɻ50GH5Ρlp )^rACTaۛJAzjI ܄e_z<NRߤE9!3k_c{AvTs̟[bf 1]ݡOoG1<_',ËbռׁZ. PXn0ݔ6h 'w"X [}1f{-KJaۥu|n- d:3LDJyx0EF[QѪF#דbM"Ts&c)]PiWe(O^܂&¬e['<!]:Fp8<,f!L(DИs#ZӅ9Kke)Y$7[y}IfgUgMGmxv|H-ȻYّfnAC_+REʢfЇj-z\nce1/iFuy,BDbb$ӋLxjYN6 |%t+L}/X)a.AӊQ-˰ӪWs izNE$2!J$ oͱqfORy 12BG9L%XIDxLU$V i $sF&䱃2 ~OeL= Cۣql(KC_vo9 P-,P-~.T=^ݱPb4Zhe[,Z+;C'h (OA gKx'-h)6XD2=e 蒕8i 9|4q6kfYY=KW}X$@xXֶHOGіGmٟulUIO6SSR%՟OnaB7nQp "ƒwY]/F%@d9\ޭ]sR[Q`tZMFK"@6f!tjgc-\@G(.a?z8nW7A'ܟ>.Y :$:@O?EPpA'~AzT8݅k>pL?+s;=v=IPjxL%.*}0_@@YP-Jdt=XC8z9k V=-?LsLeq"(r ,I*=r g NsIK <G&-}՜ ռ,[N,,'xkFּti4o ' 4EKTHS&10%YPׄ^4 B?r c?{۶B_EO8⮗n/-r '[ɵEdzYvI1#Eܙ 6U9b46*afF TR9ɴ51ͬeq,Xs*`OaXVJhr  UzFoBQa`{ D9$o-By*ΌB  6rEhch+~O&(DcR*A":3RRA#Xnt XHC$U.m fE"״ nt&-J8NLqň#9UJ,g.M5a4hÊQ+-WsQR3[1&vIl௘7_6):y[2hhPh6}E%.VO~v@xvܤ,dڝPQR]jײ\!=n:LVK49YкBwSBs陋z?i \H+ {Z6`u}z4IEKupO?&u_{F-oʃ*7u|YDi;ZӾAxfoCB^UI9"T1"1 ܘKYXR5[ Sn!$(ZEF!)Lm .XikE3pyYB%9QP-Ht12OwR #ͨ6%њ}:e13N⥪X@cV>X`nC:T]ɥP@Xhe`&)A@9%bihJ4TgDHbN8J_|Rϧ;)ך}UE&3cn/ 1qƕf`,ɻBkqx!bv^~]pTFijpk?t^ïNōw*nSq㝊SxF})'9|sHLJsed&nLҌ;`Du\zoE5:4I.EeDM_o r[Zm|QXbٹC PQngFU%ܩRܜe<$MENU*Rn2bETRJ5(ȅ)L,%ZIDzZ|UT)_)e*eJ`%tᖒV_2U&@%!XdS*7,q%ſ=E=/y;LCL,1UbJM?Lj(W,V+,JV}oS,mF ^ƚ`$1o+wT<xpcZACfGgT0fTưEBTsnab+9h0+t S$sT̆C#/{pL-U3Zp=il"&@[ShoտG6nxd&v TLKLfAitGm%Z:w;Sr_MAf)E6K bʸ:C$VAT[yi @WM ` /{3jo8 ')U¯ Z-`OlL!DQ0!`nƣi)MvŊ|?bv6-:Ad/|kOus9Mm>{n.f8Y:8tRvw3᫫J.Ə歲7 ҽnA4cr)[0S$4I_pv p 0պ,d3)B޵B <ƨliծdt=]-{xr45°`O^. "&R˟WʡV羲8?XωP~IP$)?d]kx?Y2V)#XfI ƒ4j#Q4^V&IA&xjבBƳi:[ڮ4C agcnwr+۹ \!k̳Aqz<s|ִdʜn,!lYa84_†ߓ9~2!Lb)H6"WT\r΂Y~4T~1r請LX7) tVUL$ m;{(8 Aeήybps -%.o۷qJ-B:{+FOsnf7) J6?gw=5C@O6Ya-Mpsw~[H41:bYL`nK NOwh9*F8Q6 SA}#&Am!tR .^2G9Aͤ6SOj,(Rt*F5"A|ys VSDa)crJCW$U^P8aǧ Tp;I: WVۗ#KbJUt$Żg\VTPޮM+Vz PWo.w욊̷?.XLaOf)Xּ|JyNDƵ4VL(fb%DD 89JftMr~L]0.SQ~ꨱ`π G K}kVxt,*$E(EÅo p?yoe3\&MyؒM,nGa܏m/l &{sɟzw|rr$䕋hLRZ-xR-匿lZJ-vKkj?\}uZ5W b(wԙh 6+G.gE4@8H28\1"1 aUΉvnG0}‘볌ʾ-*rkk$wSTxQ;CO a}U׈Uډ`(;mxy!lvzU%5ܼUi iikNt7iN xj Ӧ `j*J*֐rp2$Y}Fy"gnas Ht(H'!L e4@2!c[S)0VTAgsUcYS&X$Tw'o7`SNsb5b\@bÓq 8`<ŗ!bHÇ'VT(tPy-WW:RVʮw BT?/•ʍa,ɻ^()4ܿԓu6{d:t[/ʍOʍOܔ(`'Q )ed@$9g$ Ue$܇ɓ=5&诋-v.}E؉[٧o[v|E:De8tp0'I!*3tƃ3{HF:-nLAfz{;lAzxOw0F&I었p5HA-*+*L _3ÂK8цj)3g@%Q9Bmʤة uxі`x 3I%ſ=Lzz) D7"I-L$CqsLq:5"R̊LIck~2_T+g,ueaUn2RLNM+dTe6B !1ƥT2WFnQ_æ]NJ90j'I;qfh~3E| Ϸ,I61B= W^I7rxISsp!}Df"|Mb a6Bsك_$]? dnQwF 07ÅJ?;"rڍU?NJӔT,`-BBxZ 6 ("j ן)|WFqP1s,C|8OZ!PGجyguӇɡ~0?OƓʿ~q%* [5:P0s%gZ6·VLPd[?cޅWtV[XZK}]ٷd2"9I bTuld&k[C9WEj\ !H|g Msuw ;)ѷ9v5iQ8ҙCFu{ A. u4DkL%u<)rZCEr$!\Na/7KǛ62iF5ׅx5a&LALОV瀒 +s\0 p!*!9L I-f’ưN!h >v%'mrqCޓ \1w7rǂ1{q 1mTj&3w;CaB_\B5rHwvHs cTm!G~q/R9ں2@Lia:-V"1HBF^)^`kXo:ѓCE֪8 _)NIͨF %  /w{Z< c RXL쓻k8+K;/D= Z4HbuJR!R8kphċbRV*̗l#7ddr7aXƉ:XQ8a̰nҰn*eKP Aḭ,aH- |pZAڷY85M' \ "3 W4<^siaˠy(7j7 Ѱ%]>]O aבH4MTYj7`y!:AԳwJ^ߥߟ8qt?~US~R#ŒPnU[% kam5Nh” C8=,ƫ"^YFp$%oMޭ25PXKQpdJBB,0,B(oQTxKpXIHM} ^>s3OöX2fz |ɚ!ܻͮI1g%5px.-`X 1`1 0U4L#4Pc(44!*&`) RaΉɴU!vIG-1PrmI y*怵5Tc IzF\8e a\w:]夀śҹ+w,I0d#AWj *kxuX_>Q29O0X&a)ˑR|g %`V*ߗm͓|UslW^pJ1%"&OR[H]pP<a=浰?4 FMbYa`@e7@q̡"PʎBG>Qå^T"@x?.Zum+50'9 U|?bHGL( `eb3;aTFrZhBDY 1 LK jGEiXwW屲^ KnkILs4D>G<(8Yc-2+ޮř$MbPh۝Ylq}?|BhGq`"UqZ\{x?wQ)Znٰ qͺ {P~> dpY9pZܯaaLR~T9PK{|R%iM/|ӿt0i$ϊ&7̀ːRunޓ|>B'&$5/CJ Aw *w`8r(}LͼY* o{t=̗8[R+_ku 2;8'HȵA{J ҞP! PX8  <{AZf~Lz/ 1ﯚSKj՚sS*? azw_S„f0aZ ?e c+Ou/rXI;t_ngnQ~fŲOH(H6AֵZo޻ۧ??ɞYV7wqnܲ*! hM1.HujtNxj]f. ]@Ly :`L^k#:&:н?Wn"6Bj_˰fovŚOBn|n^}?>mS fpb)9g+gNɘ/QF";GP-awOrDl[śN",J=2ftd9ҰoMq^!D:cZ\=δ 0~jDCA~tyeݰBs, #aΌ>JVJ+Gkt>/(Gţ}Sz'xH"!GeIϞ|ŠuJT3'"[]?xXa!V'p$vV+ŏSY=i BE4,` mqa ~qjk +a#ayl 40 40#5 l+74vx.LI%Pp(o4D8#wPA,)'nFwMOihq@aG'LX1{E_Ĥ9(Q[O WWTrgԜ0ΛH[fq9pP VR 5*O6>j[-Fy([U-P*z 5/{`1)/=NyqXƻrEYlzZw?Ѹ`AB*D gN͕ګV9IV}GW_~-dc Y.K}F}^7)rw'rORSRq\qul'q~7"٘lϼtBJ0'5o[,?k$NF0Q{ZNt!`-i+(.h"\a J$)Nm6+5u-KEDH.N0aAZ WTcOUq蝞U+_z_AF./-dXs- ЛB T1aEuGܟ8*KQb:ZpdG2kq*AGP ~0ċkpzAB230Ga椬9 GjA  BXTgc"k}n(eF!yn2[|,O_s?yJwˉn~jW˝Ձ뿝-&E|=gDqOT̩jj(d竿lEbX%}O̙W.cFĸB_G;w\7dV`cA`c!K Z = bQP9xEW?|( ~z倫Ov+sZ"N7^j2:'6Mbv|&M)}28oq1W㼪I{aeF[ϙB{$v0sLƖ 05L C&tnCY&}D#(py̷{HKɽD"4sjfĻ0, x%d>,hX)`L"nͽHԞ.Ok +$Ldy{<v5I`DuAA1 "~d.h'<,904(;JK0 -!5$&Kr$ 3zΑ!a%Ȫ`B?)* '4DRErt?]YoF+_ڗpo'enn%nɐN9d7%k!U,r hۢXUwZNȘR#b-YGQH F0Jˆ EwL0ҐB{ezݠs)ۿu$BÓao]躗Ah];3$Е9^~Dr=_'ЋAKR!/wZ!fXk@7BI:JlwS(V 5K;dwI!FDs0{Oi2]V>?"4o(@@X%G֊D8%QlJ|.*Dae :](H0Tm}W kȔ>J<lç!J‡t7>M0GJRʀ8=aK [Pn 9uEM _śsgO l7kbV.G=֬R׏ 1-hl $I>?ݪ町 R1$u@bb(JC!e҉ j($9u>xihC _XõW cs YT2 c&xtZR.ag8 -sY<ݏ3Eb!JZJ´9$h9QFm筳-l|ַKRs\@1G<@X!FpŌrr Ԫ=KWGq*|,{[3\!ϥ +nl 8+10`RctmO~Y/g=)*n f7`Uz^q )L0JOBVqgCwow{ FA Kg2 B):rj40eX|80d>BC^hd1 $[.D%OՀ׿z0 R;bLF=8@GfEc. 7_% p\8y\VMAG-9io8*4 `..gA Vy')p,[,"JH-J$*G"U>J"P9G,rg]tS_DQ̰Y)8F:R`gfխ3l[RT _mYk6}>Z6/V埯I|UJzJjorn ?~6r88Y<,Z]۫Jٰs::WRbv}fDq<H<nhJޠԑJZ)?V(M0-.5J}!Hglh_O /ZlATU[Ⱦ H:z3F5@ -V 1S"BXbg_r|g_rpg_rdeu\ ڋ5kTnؓ`PA biTjw)A/K/V=;bv1:+{B˜J>ۀZKuBA[e^' .gÚ =YhNk10~Z# A?\繀g\|j: ~?va:^|\,-b/MRo~Q7+{QBd'K.o4̕zm\V-ý!N ,@zA[// ,fzt_m!P| UۚNzl?}=n7軻=|zW{'c$6ЇWY#x^Ԁ4bax|xlfo[։O>NՅ:׃{̷m?9qw(a8v<ߏCё~x4ow;vSZtIjvȾMW6HjzX뼠ӳ^tTSMGדdٶ"n{m*Tz|?Qe4r*@6M-ุȋaY= oyjv2k=zWj (0=x$p\ rOm)g4~>+. {j) qA\=׶/4ȕKֶpP?yc1\p=3G!lζfۼ~A&$nvZ_xr?}8~AC<8GI8#;J۽PR4\_1Z=BѪ~88ol(C pUVΒ}uJ47| lC$['lgGjh3d[A~Mz{]@/Ɍx`vpse|/}[uVBJާeۣVJz!Rsqdkh谺5 F]иFm.8G>00{:uW-J:KV힕Sv5bfNdGˎVϞ;G&IۛoZi9cmk\IF=$r.W+N]^oyw)7WϑeH 17[<|Y9es-Fv{>G68f;'/:_zi݃cit _8z֝9,NX3{t#']\FڇUMG0Fhu0q;x4(IY!iv tPV0JF29PP# EPt\!/58:>.% nntw~?+41] OY9vsJ `:=8 )XXV@oϤuKV ֻ`(' v$&4a4^Rx;qh wri^:w?9q'TeTbIXs=tR3T|l/\ as!9 LT\ML]6FΪGݧW8I8:K}G 8^R ߇L̩/=)ǫ'1w}vBs~SteK1!(m+ tY-arj!m.fN/bOIr9">C8Kc(Qƒt|ֳ;ZqeOFn;+|PL%c|F|B׋ Oi#ŷŷŷŷU/c:{VؠDw@B#bxC]P}@HqʅqQ1w1gsCћLߎ/4jf10`K? ek ,-߅J M|r[!jJ& wVQ]dQdݛ >1 Qɫf~)X$?c2_os_t ͛1Q-dLh AĬ}CW VƵJzmyB6B7(y=bO( Z[W/f`8uf}͛>bm֖-yz\}zd>*l%<壎|^?cKƑ֔$Re%m1QA7.=H+F՟IU[8b_Cx'#WWVB.yɑ[Tx-=Q&cÞZc,JSXLjg?vW0 N*Gn82W?'44lBFr@,bQF "/mD:cB#-`Mc$Ek!/Br#6Z _8$9Y I He~XUG '* /BML Vqvp(# Q ]0Ųh6W\Lxɼ"W':<7W3µfymo#5圊<}&`4 D B@j4xYCYGK]YsG+x }(3fbc!e_htM.|zۭ|{յp5 \)־9踟-C3 #?%; Kq|狠s/w߽Y5sC Ӽ@3sj l7_xd:[.ܷp0zۅOqwzluY`G#x> pr4$Q8Q=L&e8Fʹ3 @K.{x$m卨협 ]a )64 MI땂 BC3sO:,$-a7ȼeMS:#P |r+8|7#-tY-|:a>+R0lnBIn~vfD~V= n&q>M9cdv?V!.#y1Jm>-M25o+2~} 5CE%9 W`Z3meȄJ$rёn{[^0]O@{ h6`%}L$Ye%wLM@%t앤{R^O] + K)a.GH-c>iJ̈́)9g6D" P/Sv6i;.')=?Mׇh` d>՝>pU& k>ylc0ſd?| l:u,&72Ƅe(PՁ̒B FfFW9 YHh(ބgҒ0r&U3&[`wrn'B@- *ԕ-O3Msoe,yԋdr/Dj-zCמJK"AMӷΓ+$y2zBb].zHU|n#^c[ޗY奢0,#h 7`='F-dA8?ؘ?qzU4g' bJvq2ЇQ0⨁؞pwJJ/qj\.E\v0s40Lb~q}e<ǩT?9d0g(hCq8SEXFE1L Xl'@?UyxCxs=$*saf:arA$yZ+1Š)n4FZ:bv݋q[".Y@Ћ8t#QћO""+2N&"i|Kˏ/.v~I>N^9K8^t %QzËZ<$<` Ӡ;Q|6uYSc 4Zݼ'SZ͔GwE'`B gig0J/erL<gWerk|e=FT4,>Je2ZX1;xq|vd,xP=: Ģw33L@~Mo6xSN-RrVBBe4}_J0|\4ޏA\vܻėH<9ç3?DXfOAӇ *蹧a>| WDNxHɟ_} ) *C+qVv<á娻7ߥz`fܤ %Wהיi=ةM뇖_ѹ^Rz/߭د .4Ñހf T!rIb zEJk&TDPR2ssgbY .m|Ttn\f2^w,Ko\NhaAg!iGm3yC`Ǒm6 W)t)l[6`^i6Mf8 _/ S~90fik TL1-$\*2˴),le6kQ0AOX/zJ}\2"^"o JngL2@_KP֔>A( (aNÐ=9w@ǽp7 $Ml$S~Ri&]S7hIVD6` Kn1i3 ;' a] پ& %F (A!C`(@K^Kidh:RJP_zMLy.(`{PNfu T܆1QJK-ի tT̝`ZOy(~[ BUzJky}6uaU?\fIP΃PUr0C?ˬ>̖uvt0/U)xxFa \z3ͽ&H؛կrF`%P 2o.35zl2HΝ0[ag'ᗘ#T'IOh!:]%B#G~Mz2I9jl;RY֊IůjaU:ezF=LKRgjB6_9`> lm8p!?m%ypLך !G4RO-!o=)òu L}<.(F]:5MPV*ًkE$)qI9,z),W!ky G֤p])-WX\t{53T:l\ƑUd 4s#ODl+ytEE^  0Ogw$M`cgy:ɜݎ6Ka^p+׌LD s"4:v"򪧜9Ư&.ه':A@WXoZ7Wr{soY49\qT'ymLY##Yc$Z.MScJV!,sfS'!$I"j:(C㗱n!:3mpH5.tvݚ'Bqëuŝhv 6:NFC>jJ%iqPqhW(HUwu;NwJ;躵B ԕ&J=Ƹ`WU+\2l8$uBm44mN^0,:\d-鮫4-˓_ur"ߤN[CPk,ewUޯSy~*/[6&qFPRb,,93DQN!z2'\ukjodUX!j0$?z%n )Cɪ*@~n醂ܟML!7MPM|EI|*Vw:rͪ[0FοcWǺ4o$00 |{]`-1y;§8˿;ȿ th0+xI<ol P~3]H Ldwa)Jԩ'*E'諑lYT.Ơ96%4DaA[xyZ ʗz/AY'Q3AIM߃ nZ`vkۉ2~_jۇX@})NPcd -Q+kbe\ m氳%4MS J?{6MQƣJ\ںNM [P{od [%up#,1)Fk*422 w:{ ʼ jP##'`0AȘ- ~ KA7,zCA,SASa)h!#Ʋ$Ob\] jS4W1t0$C>2 g&2*R+MB8%QfR&E>Bx_l$FuW#x#ON!.*7&vQ F v`܅KEdðZNOЭj(Ӿv(JTL׻ 4+)YaoLIz.zV038@Gуkŝ)fKOjF:Tn\<-l5p&zXPiG"7oyMRn2\ȉ,~O>}X/}r9Qz5*NL0 "n"6 1ZS;w5wz5*cT=rhWrqr2S>5h,;']!C}=t  #Y5fb(Nk_huZѿ>xxޞכKWrUy̍Ds.y8OLAe-PiS!m#zzK.;,꒱ z/w7\w[cgN)DJ" r}%5 5~"M̓XL:/,˒:؂-g9{zMbK~3##|͵y~{g舚!M<•đ)~[W&"sM68~iT:0(ѹZ9UjIkrIVR,sS%OpK;=PY'Vg$q SqHB8s\Mh3oƄƼ՞NMPjM1_#ƛ]<_8]_9.~xb-:W>ʻ1do@b:PzROMٛ[&`wNJ3Diu8nڠ$ v^h? F{&h+N~ vn{#vzw6]q?ygV6TKwyAR ;TRvGܵ)?[cFZ:u<ʋy.]]':Fj-"ş40n}_ǯexKھcE+7x6#thP&-YUo[ TWk+Lf{_ᚧvYb媽QGeJd-P(tJ7ݿ;{R'NMwZ-NE_Om4n4}(Qv=oz$N.֯^+V w <' {HarNjNpvkm /ov y/T\5x3"$e~IÇvrlR-%3%a*]Tߧt@7H+bdsIIkB)vG''z!rh'Y1@1"`ҜG袨Ӕ*g%?~2j&8'=@#xH|yl4u>4̠tG%y ;3xO9!X B EsPR+O٣\RH?u0AoF9D\skh0}vyGF&Zok@KeДJ󿣿=CBu7=^'xr^QE7b(8sy7Q+WN!B^r=B:E"t,Y"͉IUxFmۄ2t…IL81rϏS^?_޺LEۥm_jnaϫ ei3]g_^)bB8X{ֳ ٙDX&5RĊ& Rk-Sehl"IF%­ǙVBYV=q@3UX}ܡm+%~R2!/Wb@d(,ɦ< ަ_܉O먉cUӣ~H|kAU-{f*意(x_t}X|sw9vFEe_ 1*r-rO*ݝ놙G7E5!)ӆh/ &>Up0Ahe@h[MkpL }D9^E`gLS4T4j(_ڰ}ip%$/4Ȇ=($T[ .0(PEAkH" &eRB܁aX7#0jTD/m[H$ykGL^~la 81>c?|J}>>c^uZUF>ݑo]/YKg~͟^&;^ 3g$n@\ ^aR1>=v7"U4a_h 0$7VH_l&!fH[\p=(o+Z:v>fw~oߍtW'Ζ ɬtlA2E8w ]ǜb0ի?91fڝ^h5 ;$hWlYp#AɪMVϊˬK'~w#Uhێ]Rog;Ƀ=گܪtgs{tVHC2i25_W<'vd4JM4]ڦTSs7걈gc|Fث.EnOymn3bgPr3RPBNWGMc*x]Ԑ g )B3dHWAW ~-CNԦ^+W[xU-,_*%p+#ILy*$d*m@r.HH |/jVJ}yM>r@Đnn:a,6tBxMṘ]q3 6"Yn|2kB$h _}!-JIJ9,Z TC۔䙵'S\ȎooLUf9s5wmz TW_'c$O^}]X7K=C8(iá+Q3]_U׭Eh0A(WrJ"3J5k*.l0#w]#U4\6q;M^wI)vnްelA&uЧ> ˾#b6O.fcd^!@ | B rHḭ]=uZAScr,NS K"F:L.lLbI a,)h; A#I`,?2YSN[iBy`(`IeˠL :<&,{B , eZ,a)0jКaR{=mV}T6Q+ryjǨǂ>0F ρ?8+62Cc{o Ǘ]pD aLJuqMrh_GP*TYA2K7.&U!Q%!) 琢>!J+PUg F%+>h($RzH8#" t]W̌v9c57OWƂpέ|%V {Y$Idu?Z0nx!KNSLQ`UŽjRf J46qbw`yYn"69 e0g+evAkX)X4do t:2bl^~Ri[iy3qq&ǜ0Y(,`Q%:K"}rbbw6jG"qeKAU"x8C cux9M ("} BצECր i7˩W6ǁBIf,nQC%vnm){-oR7\Q0xbz%?UJ Iӌ[MJ^-DvN3`i]iu,3zӞ".lYVkBc8@JAg3r\#dyIzLi|mJk0tJݎuٶג2Rc*>D'M2l-"^ QOp*jõѺz0BS@`wʨLK2199P3*#끘|SRYRfITbaG!փhga$$l 2Y!YLS7MFS, /WV]PGo]: E>֢X3JEd-Ij5ŌS 1˚bQ5@.y(b\a4@V' RO1f33կ(|YQ8%lqd7sVYdKX+; 3%P%Ǚ&5X!EppM:tT`sY-zXu&@OvDN k[̱G9cbs AbP_wЮ?RO@4DʰC^^[f,yMN/R4AuK|.뮶c{28av8y=9-&s$iwٷ%%W6Лo~烀W΍.1`IQ*4י/Ω4֤M',X!MxǗEUe}j&mRvO-1x\d07l j^# /S%k.rp a%]dSblޓUK"YB""G 1s ؘ%H?Hzթ"\ȷ'ٟN>o+w_I˝,xa.W˭j w~5R!-F,"$逿$SkK6Gj=g*Y\ C=wgImݔ:)(-j+ BQ+zwt'CmgjzPI=Ժ|;;%A[dJg?rwn*-"!Rɝx4nx @: Lq{m۱ld%9Tt>U>h[gt!;!ln?gV1${G`hw+8H9j|A`g 2*~cO/ui6w>Ӄ\\eL?xzT`N?-r#GGEڶ4jsWRG/h\^4k/kuO? G5;k;x]Rߛ1lF)"8T0P ?lؗ*B*d&XoI$AHZN 7 AD/1k=>Z~0j:`aqo2⽉L彠O~w= 4WUvvu;{+OfwGG TѼ𳪲tf`qv::ۛ)4nrY&2}~?d)mTQ y_v?@۱ XckXlI .Olp9 {GXl67̙}f0u]÷oZ0zdh%\zDWFÉhAM21p!E1)NRw/fX u'?| t /n[]Y2l﬜hd,%|4 (|gonN?^ "Aƀ/?hWLi~98>t'mOZ|k9yh]R=ՐEߣF[EWXRSjyQ‘fR3!abMF!C UAOr>sZ{/77>QN~o_cD?W+=.[G{򦄟`ǺqCv=0MUiѺ=57LܭS|y6qƽvTG=s68n ۠}R(u.Ef1&ibu1H(B 󍺡/tn˷x#65+\w3>^_䋰D' @.1ےS"Ab"o^ f 9JPZJ ' EA8 c{PێP۩AVB{Ձ,Z Jy-bI2*DS MDG?{A 눳ւz6Ա 1+vP(묉t5,%1XH,f^#jЯ[#.jg{aCwq8tX4{f n|x`4GϜH@ н_ =7B&/jt3>PW{kqGm40I:a}oQǝ?wn^Ǝ(emd`ýy7Hr [3ו(`Gc^Rm$3np&z5-݃ӋKroܬ:S78+U޲cWr S*>F]^&2q~?Y?|o׉X𾃛:=s!WggC8cwOlB=Y>; gx9n%vx'٤Y>iFާfίgmO hkqsla{Èwy=ڴFC 3AaD>} Cʹx\n[v)j5㍛csj"7.E@)ɾ&Cߝ6FT؝p$.ޥ`ዉx%ɃRǨL)I˩8@%eQ+ٱBԒ4j+F  0XH"`d)z,Z!JVNc5>OgEu juo di[ٮ6HW?QJ=(`,gA8}L.Pft{mEꡀ6zkMEIt>l2ʘN) }q kg:aX#{^HM3lxjE&Us,Ѿ|c)oEJny$\ldb;PȭHZfOnfg 4Q}K@b(ֈ׏ai)M~aD8{>=m`En-N闓0\RZLT؄d&'ҕ!2'Ѵ.-s[ҷ=yJe9ңQBrՉcpgϊ[kCx/p:muѓbfrFoCHGat hqoY{sꌊhK ĭ8)wmHcvk&|?eLpIgvp 6m˖d[DFenǑ$&bX mIkN5D,[qf>ҭ18NVHKyct2>ldTx$&VT:kN 9ʻKk˹O Q#E8-0H؂jv8GkXYs! 󦓃L2T($B!ðFDSh-3f U&c=T5`.Re#n,C(,2kZ8њ1V07W`>I aVMa#i+r4w:lr_X/X@a) 3pRO*΄ٺ4E*B'tlMUaF_RtVFZЦ7 `ɩE鸍Hp7.&a1().t>2k >mmH3s&Q[]n=cǏ(4a v.aIғ{՜E%#ߗ$8T{Vh(hW Ee79q1$`0B2xIA~<, D$\*r+ȤŰ ,t&pM2}]s_A4WA{tJL"-.72SL9q ۛ k֜.ؼKuD;]~~qYHW!Pv+MF:kKu@aFe\8mbC[\r% B2aE5)B"w,Lm挛OF9 #K]]yd 2 xEƋBj4G겍`:'Ql'(v!!.Mœy wqq\uV6V?L]%-ԓmƨB J-{^QUWЄm(:tЬ:)^2 CU}kG9O( b$#?NGhH[iQ-SQcf__a";,fLI)Q#cDIs˴3a^qc}ԋÕW}ѮA`A];/mGG MR>yhb4b> R*#4Lې.Te5Z)Vͭ *Z4h+lc/&AplS*]@vZ}RwX*xlL8ցk)fV~+kSw6߀p¤TG:;M%USƉbjL[8?VM̳Y&vM'BbEqNcKIؾz5RR-tҤOg 35_YRIXȐO-w.$. %+ٲmYA[w37zδ?Ǫ&(ׂ6hQ&6TA :zi$V:"4vZY'c v}SҨ[~,{k\w`CFHN ϙe't6C9y,#_(b´>Q [b8?+h+h3A tKhx)6H$8Lolg\e[f`6:Gnx1u@W5 Ĝ@(- QZ&(aK&895`;FynM a*F"e$tf[-o\ dPp$Z8~9 G<$,dtܷ҃;X%eI VIX ++g7ʳfZ*W>ܙFLw.{>}qWYT$ |n(!6D]>ϔ 2 0lܹĽk޹ذDQN _rգF '[HI24+r,i9 ar 3؀o`cXW[mCWí+Rs?ǔab@3)|/Ũl%9~gLr-jJJ{].wוa`;QIȂ85 ~->l$3:#?:)oT';qT1`ɓa98SARATThC;Fv$Vܦ HI.qMBQL-X0r!m[%y[m`bCL㐳ngwhdN`eVo6d=H\Fa얹H]0bEv-Cmw\Hޥ8FNϧñ=|rRz̍@GTG-0*z *2!Uб+v΅4f)%5)5Zwt:7:dbSJ.in+ Hg_j ï>-Ó0h@t8pK]LlKTwg;yRqc dp236w9vKMM.(Er5^[ ":2JŲOk}W FD/,+;qk;Pw0icT /FV7;K̎zz _3 ( h-ؘP@(Mg}NH{&f't_R2= sq}x{2D 7!]e5r-a~ZNU九Yr)fNeVZ   T94ػG?^ @L,O) sDyHnk`XB'8RXd9?N?6PQn#t69r Y'tc_cM·xcG·{NN@+y>^VϛG޸?w⡽zoo}׽MNG`5o_uy t7/^ӟ^}_^߃5N?ųw߼}O{ok>W/E(ILgW߻!w_pH=`d/>Lj =|O| egf rwn*5ga `jWkl/^^Qg `c6_mgG'vX|M|M..[+.ULNB\ t,7L+P_›O_u8/'fq/o߾9XԅɸfWO{Ngw+['yj>{}_>χApi O/f7'Sz ,wp@Hӡػ6c*~P,ťu˃^: Ecy]|4c*;tKߍUz3B1+BUپ#~F.=n0zy^߼}O|w:aK/]~{^g/ysft~4}g_Iؔ>Gf_XkY?]ˏR˫6tY}\vK e?@[WdzW(^O'J_qA~8/bP7fq|5>˘BNPقNʅLW&TWf'K0ޟMD2貢jJx)I)6c&so}J~a`'5|B(&^㊑Kķb?w9^MaK/ _qZW/`RFX\ҕ^35`y'X 'yIuef*RcqD36{kEF/E 9avT/jU3WpAxu%|1zK='I=q{OT{OT9c6[;M $`A/8XmfEVu_(j."N2xc$I$YR"^ipN EposS!(|WE( I$٧ I% tA.HI $ۂ-T:JyH N m7]`|ҵĥoYhԮn.m,ʅ3ͭvyTkŗW4ҮFQr:7[6{T$̰}l_ǠHe۵ ;.6>](?T$7ڊ4i\d! QTrIRAЛ X-'RJLWW׿>ջzW¸_~iUJa]Ht"Qkkj%ە,Acԑ~ zNVx =`bL:p  EJQjƽiFsX>S<}p5Oh#VĂjpCйhTTi LkW)ʔI!J<3`!RBHGI1ІmdDG &]"%]"%vCڶ}Z*!|YMB:'#W@nV$݇N9swJWeEv ., JϢҔՂM`b I=B1#FrtÕCF^Nc`AIjgg,&\D\ G.- V2&[Cْ@N+P(hiD+xMη"b#iպiujPXݑGnHBͩp:MRx;r/L|>[#!{McW! ʍu3^衃 5EfJ[ 4%hGiT[EXeK֭I)H :`>{R/{Lxr;@g1/7 N`#Kכ]ެUxsc 7+'㚫ˣ4n]___B?^]Ca2z ;]'E5!ELח1Z Q*(|9jC4Q6syˆST1ηH-DVB0 -rۉPb!)Nf9/}e8?kZB/o#8(/"7&ؿ>ZbI 0Y&K9KFi\._R͎@:-tb6Đd2v2x"ӹDh(BILyo?"םLMz4{qc(^7sM<gn5<;6ݞ. \Y%jRi&rSW|j8:-e1'|S/]~bLeXsS>Q`zGs̞/=^#2I0竞|?~ѽ'FLGEBϝdID ʹd.RZD1$^,[O'=((+q({†p//L~YIT\g7ƅvs6W)ll21@0=pP0XS6CwˏUAý:/"]._lR Im7o_4:i>#i3֐uZS]FcLYڞ8*¤iT!e+3Q8َA6gVvgX~֌MӮlOdO,ٚ%߻> aԑ"C43wŀbFV ,Sn L}dJ"LjsfY4DzXK[C$FzF-BF"$ErJA b&|*򁊬pIav+-hzF(%>%8UK(t .% GN.O(< Í'V4:<2]* 7=l}5VTŖ7b %;U; |yܰ9Usg}=UxڰN Vʀ\W;V6۰,Am& &лd\T1RH{EuhD$! L[J=ꈷ8lI %o6] S] yڻ@{h]}}|48 >*Rw%'FQ($&m&1;*!JTYPCd,I g>{()\#pYe Qu HNn 9t^PDIw^Pu^Pu^ei?hFJ:B T(D$UFtzIk H5jF\P m h*MvJt-ch]V@L+̋,N\l֢0Ÿeb\޵q+ٿbӽ\>U b]\ @ͧ"R,˶4q/ȯ̴\'X,VS,ކk%1dx"m2?S.Ix`FϯFJ-(˶y{QZpE*.DcԜ<#s>΄>rլ u}RA>HO'uNar.4*thN`kELCx?QΡ/t{4eLF(V?rO?vaan,Ə~+.o3|o1_=_ר [۷X|Ew9uzK^__ޤ&߿8׳{ہ=%_IB68?~}QPIx狸x'Kt!B%\Uܴ/hrTM̖Ιu l7<6ȸQ h٣[~ 4{gқݰɆOnޮ,{@*0SAI e5p8Tzs>Tjx%LSLSHm3t0Kj%2ލZ; rMgoY۟n}z{ѯ9-I?_ÍddzqLC4gUmUP&bJ>I1qf 4aBa#okQq_DG&Awf*Gj Ԥx.?1К g=lB R 0N1A( UMUL0d|p56ҸX]&vj},3pVͳx}KQs<{I:7 j*#ͽt 43d^Uy'*PȮE-d5hU/EB2vT Tl},Œ5Pspi68hBބ;h! 2& U[YE3ܕO-'&2U1ZSN׹> 44%>S0$~he9bɨ-gD$U>BnM I[x}Djo<6T +p}5F:Q#kIvJB䫶_*324E冘>JIAI _qPԴm1~MHY:ik;&@e@)GE{/J󍽼Sb'\m˸Wx7檯vǗ] o?z{  pW_n8 ;wig&戴݇^_={_ڿk@7 =j7767 J"ɽ c*/^Yw2l mHw%5G t؜=zf맫+4ջ5\.oYoO x6g]Xzrom;&Yg1SrT:k@9dO.{v|V,5745ƉAlV`J ٘ILZSJn -f)Ad1%cNo]b_/W'<\) _!¡`4$!Cm칧!. 7hI@- Dm(TlY;1atzwJ89 [ZĹǨoHsn;f1NbGl؊e~2Ut@{ytERhU eGhH0Tm4鷣#F X򷗈jC3ʂ)q;Dekձ' T#ke"\F:BӕųHVjP($uI<9mN$%'J+o|Ӯcʆ@3&YeF O~|{P#?g“#Bw=l݅p iN,FCjV1,cA4/ z b8 !<< HB wn>Ild$ɛ@٢٘|kL:DQۗ&-Ǐc# u`R@VƤ0}ȋqxP0 Zz k'Ͳ#~ 7iBs嘄H@ޭɅdw]l[Q:R`GESvtdFD螕}9ЌՖA[ێ ᬂ#rT'ҏ0fmWwPhgQ4fH }p{g pSQ It J lq(¡7}QHׁvY %5om;ҦSr DLrAZ uVH/>wC8m]o+'c4:ز1oks@.ZbY{v_F;m= Զ=2⼛ i9>{ ɷkU0ND#7}} +=M/Xy^Flj'-JNTARtej.YZɅo t9 Pe0Í4s+]4l5˻W)xEH~1)b[jB?rWT s/ Ӻ&՝*jd\BjZɢĶiJ4VQ5SUCx$U'q_d#ze:1Cfǻ^g20\>0{4DiRhkIZ9&6 ,ФV!*!M+H 1 FI^}:[۸+]K3"؟(,cJg\l\DR@ϥ5` yoۑ6lΎ?Ueò1DJ}&}_,+[V'g,+yG0u{&AfQx$oM; I*π$뮩O%1iWxB3}ߋ>xw +ʒ 1 Fpjzsw?a뵛$vMOX= Vo,EV`qGJ}`6ׯo(MLh ˥dcMY02q %bR^4zݯ &CPDfڪk),y ,[ZHa2.t) g.q]E޵pQlM65.쀠@h Nӯ4J 賧Ooa{%>g,4Y}GyUn90e8PO bT?rByO yť0 ai^ dvkz~C$ٟ(ĩ"oR*;!;f!W Qddm!3Fq h =!H j{u7l;֦զ}H49$h aSj2宂MYj&ZJI{*ר[JX2BAXc .1N1a*se2dLcT@$&!i;X#8x6ID,rl&U+{`H-IJbTS-D\v b$>X#]l٦^ޥ^qy{G*XJ1 ׯDPPi{%B (rar30LrNّ'>p(* aa%ހfg۱6>.qI'*ƥb`U!$y%ܓ4v`86n _KyD =| dvi3(nvAj .)7uL T?)YD0U| 1ֵAmT* α3Oz.FRo.\ ?{zs9e439``њ[ ݱ~ͭiGA- K8(pVAy~}bйM2GGl41ƲM:e YIus~FD~ϟ={~I} #i'i_†[d=B}\ߟ=wW/m4Κ9ME)RC3%$Rnn*|S mXoɷ}ҿ*v/Mt0G #KlF' lm~ٻq$W~`eQoyi`eHYY3|S(KivZ Fq1, i)Qp'"1x_]/f/Mfma|sP)O`&|ݖݭ" NtIįښ :W 7|8 W/_Lk1+bv}e_3bn˝d׷c0 `4i>>^e9LPI2kR*)A]#V>+s5j "RX0V23ބX^6}mޢzmO2&̿.&\_~ D64~,T,RGsSx Cy9ȌEƝt?,Ü4֯7RUttˣut~i @o/I6u\zaHԎZ$;n^J(jhR!(a\S-p( A"+q¸ҿ$')rt(lp\wnf5 6zS=l[ΑpD]T|2IalJxiab2tpQ1/k%#?צ E6B`,y ?kKL2LJ]AW,Nv9H M4 Ҵ U.̅)$g45"s tYXgr$̓&NA:srͽ׎V( W c*50Q2F'C (p?ԹWoh% j7Zrwm w?;Z3NCF ڒ"TWKn{"%ם%)S  cw:X狓,Ck>ʴЪhʚ xڎG͓8~i7Қ2)BonV|[Q9heǍh0?x8K ڐ XYWU =G,-LDʠ5^cUVuX?LMuVaVG+k$DhiM;]X:]fk,qygk6ƺu3csmeR,j:2g#! BѱFE}H1$oJ>|Y&l(+}|L^VF+3T}x}u*K/B15u批 Lݺ'frg/fK' e+%ۥl`Cʖ/c}wwVMrA`*ǻysrV;L'; |X_`-W6Tf|` =^Q'KXk2xP&3$3JDKxg'+WmRw($ed^&C.PR:+=<0Mu')47 t=S畺Jۓ!M՜b {.쯚օ4"c cVf\Z?P~Qъ(b}X O_MdvcRQj[VeOQJ[f|'`Q-7!waZgia{P.d3vJIܬpMίLl*߿׫4~9i|ZbjgqҠ}EzS)\;q鈔suZd ŎNDL9oOB뙷3> g|4iW[0fPs !)qZ_,_g FiFgm8 ui[DQMcW]{V5 U<)els-*`&!1֕*bleܷIToVpq޹ۗҲ AfLPJEeύhTR˝o~{ƪyAuv`87%Dt˟3!D f;,&ƀ(/2$V&@TO"gc;yo*P0 h;{HSi9[JMn,і:Pr}9v-m mDĘo2/-2_v2m+QlIP1粍I&gKDgj# UΈŠ3V7@g=s1nbF EZ6X&"Y&,@Wr]][5/Ǡ *Rk5'[1;~6kQ +9ɜ נX{MC;)~^z&ke ֚HVc`x&.,9Z1\`/y-e腃SpW[<{ӆC8Yj-]@X/B]ԱI8E,\5J&1iU$&W< 08(8 ?'` f).D,|MeڀR\Ym7UQvS iK?-{-QTY%S^ZRAr^z|Kw-?q/ ͡4I%\C͞HN-g^KjtV9T "N8Jl~Q2񎾋A^ ;{EhG3MԑE7$A7^(Fyܑ~R@QnkIqĹD[tOc|AgAOjw*Fp}­#8"0|5|l^ͷ'z^w0 _zngY% R!ݐvYX%U:}6\e h.M>X2&p8YXޞ՜+W+e/-NK+!6;[BJi_b󏩈#X2'G+kui[ [S)@*RuAR/Io7xwټt^(ݟo9iȉLCNdr"jNde-3D繦2U$8ԑ-E`a9 _lwmKОii$=A’\lt(uwԅ2IYI33%c лYŤ ROGTX+O"%n0XąJڈPHT1p1t$C)|i XձkI -"W$Y\I/ ɪN 8[ƣ8HJuu+i7 Ec#D2ЎrN1}]n%CE\[Ӡ6LyPzP;SE~)icjMB0gVibY] 7h*rpa$:fA QT{y$cTdN`Xqxkxm*i'i'i'i-xl~$} 2؏"RkEDꭴ$2Vc,X ȷVr[Ɨ_| *__)~]S/ٲIAԈa}ߺULl*O|N߆6yW| U+~{QvjYK-EBytO 9 e-8tB B6 o,LWv9'[ Tj YyU0^ڝu92I1XG^Y` +* >FP8s~(؜UgF?[ʚ61[(a67t~q1HS7bYx1A2atٯ 79a\=-?wTJe#,=DGS>`E ic cf@9E)œ |5`t=B,icU T3A#bfO "6 #8ڸ }x kYA\׌xLƬ.85ؒ[Ֆ䄑[8  qd0{Ju3Aש&ٴkDT@gH Íyi8G[4~KRu&K򩻊^]ւOui"1F3ZK<1̱LA\։xv8u0kۈiF3UO?7-o**B+O=+e 5BUVXrA[(P_w>(K>lIAwEv6E.eipO!7&*Od=c&#df jhlo8-٫~N5)GlV67~[ļ֝|xGižڪU4y⪶}[|35ci < q^1R -a?Wq`!MQs_riHi?v:QQ--?n+Y"0u([o a6[Xݙ~u+ۂk۬aEAr #*kuɪ̶m1)#B]c&|_zNRu)z;"vMVjwW+ނ\I1C亖#- &ۻ%T0 ]4 "_J"7oc`,vEZ8ؕS{cW}asFcn@n~}-ǟ{+}M&ע{ 9Y)>~uf/V\P 'G8EQxAp93vWTߠwjW߬qm鰍J/ػdo_i7@(ڗ|o1TfbGW\*wU AQFOC.6녟iiߤñ~KaK" BKڷ*/۸a8Z=wu1#%EE* < |k3.Zʂtfhۨީk~tϚzM6ΌnǕW/;yOG !6LM|r4$bLFV#d% Czm{}i^F >Q c6`Zejގ2i1Z/{-8m5FcpPta,dh>UN76"̓f}orW#D$6٣ew<}2:_O~7 zg/~=(^7{GB\M6 Z//e{1h.9^tA8 {Sv)ӞϞ~N(V:3Q ^:,5hO88L6-'h|= 2Mfeu,/H?Q*N*_ms}1*E@H{ ҵt̀߆dV;*go2x O!ME3><.CΌ: \Λ_}UR$0B;eI.`!M_r$6aRځr}N.ց=܂ <łE8CgДTTekA$Uclgyxsl-31fDz_pdcXܟ+ָ*lŸ!D op߈;&r u#Ӈ.RF]0۟_Q8 ')|| =u&Aʼx܎4025&ZiBJA=IO/{lH\ٙN3p0O-\ Z9I}߫Oh:jg[';NxүC$`¯)G7{  ؊G~j%]k-ŎΓF=G`v3qsn_U?376SI{z~GܞŽx<,NƷݺn/k=tۗ_{7ݸHXB[?}3q7:OnY=ȴ­)!Rwݐ;u"oi(7~nb^E{4ӟ'wCs7m x닿76%񕝵nY5#hB iwV xT=b6W*Zw' m$65C<;oF_*q9q1"ǣ;=e/?5U*0r|G <%^tG6 ;;՞_ l^y`f&i_]i~6U8Mk҅gfhcPnFW|zv(#܆/߯$w#F PvA`0nd 3xgMHP1mv@1Hrpi!V'dM OG0\}9~;L;=~qr#.!8QV豂oSND71Z[8xvü$?7lvONK?deɜVO1l]M,ԙً 4!3i<Q.- MfBnUadn3g&s C}|}lO`S"|"$:V)lbG3Qʝ >%E>T,}|?AA'IDn xEhw (uQ]{ÇKȮ > ^-՜q{t^-WK'WRM9W +HP| [!PLk"P[C/,P%_^#e=3Mo )8@NҚ'$c2RJɊ"2&fn7LYT$yΘ9 h+t ܝ.-}/Sg|KN=7e/{BUKYXQ\  (ۚ̊MKܗ]N$eIɪI`B8!P(!g̴㖙v2ӎ[f/3'PI4E n0 ( |i- *4%0C%npBNنJ%m|\ř~ r[ɔƹLqsBffca=N.l6eB ȧ#TAhFBFPh4gP!QAi`E50r%@ \PD$SF8oe' ψda Q+E\С, c\=N 02 XF4Z"k wt@Fz4TztMKSS2 8]6VjC!@k-VR9C"}N!o$庛Jն}qLYB@"K]槬}ϽX"@['\ybcOxD"ab0DiCBe`YZ> %FD="DYe rk^ܥq[8ۙ-0f*y#x -+)Z4 4FFa( [CKM|I:4 _B  gd pJr4#BXd73!]׫CX-΁ɬ鮮HLJx|}MQU>DB d7;qI|v=9p2%DLQrFU(džjN*!@=ÝueHG$Y6"֕qR!U7\jU QmAl̔6]%5txDJ$C-B$M(aD QBQJLFjcRP C!Zy _N]2t ( m)D2KPCT;2d tHch%1a24@kfhkc.E > T Uhyގ4ˤ#i`6DuUx< mҠ13h=Lۜ[CfoE#Qo =OwۢPf}ӛVB/f_]?$&Gf/quk.¡ .a9{8tv3,^}JW'D!J hpn/Pb%2g#^"h0#v<ơkEK#FDp1]S aw$Rsl( J*՘'LQ+a5S(AG [FTYMRKS4D-8aK΍g5NBb `8jwخ3x :!)EwN98[pԾ+Ë;~Bw-&X@\FᱻN!/ҋUy:>q0D8ygJ5W n8w r0)~Ks7"Ze]7LlW</]O74|I|2/ƓShK8Sxlx=4-n0Ot÷g?DɯQo3 i)u1X͢?mŇK{dK\-T΅uHzZ_}C._BW?{w+߽<= oZ[f%Ȋ/T}~O~_Vۯe,f`9MnooNQ2]pv<\fǸrx9{Q Tb[s,o_=%y[`?MAF= }!ǃ1y*.H\r WBQz -vy%篿WD2nY5Ƅ}ҥ@ B .0l<m P#B-M(M~_mws}ijt9 y]k1J-Jnzr3v¸ڑj/As)2QV n m;L_-ٕM:Xʛ[$];[h#Qwףqu1@A1Y48j@QyNe- {xvASgDKT;?(QDQ1KahF#s +ň yiu1ߚ|(}){A KDq%Ұ.Fѓ0͜TG Ӿ^!`psK.%7\X:$r&e.P, QJWfm>I ݫ?^_e#t>M`jiyo_/6M|YLSۦ#GhR`-ӧqγفWHB &4e Xe2іږ@юJhHڣ#2fŨR%TFF9*,  B{$!$k8kx ,3f l45KK{ ZHGHն6:F Ay+Jcq h;(W Ēu #Vħ#,ERP'D3=4wg9=˹[ rE0fk`C&z&ݱ;12-w߆|#rS 9.ZkB:'AN`r9vƔ&FAtd.W24P)߾zUg)yy3)'`oYi|~w̓zjkm QͻwXW7^Qڼ< 8;eux ,sˆ?ʼRܼWKbo|aSWx *Pbg ;iWE3Gx, "m8Կ ^36Ai>"7(0S=zq }I Y)h\%%T%-{h 3M/|qX#gSp3k). &~hi`o.oR3vR WӰI#E1EGf3LGS c\?ucjRK>HB'lD7.bpwLjt!A$S?}Ծ)ץ -a?m+9Î5#h-QRs[rg]QYG]lgWdMG?;$m חU|:TYADYo@[і~6NqYD Fs8K*hR>$LSPx%BCadX|a*\01J|Z$!gCϚg4$!lI!㣥(JlD )̢NNJYt/d('bOF; ~pGgONZ*HTPD9D҄Ruٞ|BZkC;܊?7TAo?!3Mo?k(ZFjT֢1 //]2FoyCO+ 1#,̚Z~̍M/3B /IEѤuE.F< ilpChHjf -ME"Pn2lXb)dL )HϘVig486nײ:]^d5y\ 4aDɣ$7KAFsQ"M7_Y\= Ri~|ӃZ6"wΈ_^vVd+\f->S7q9f!~Sb}Xߗ|; 'ΰģYb Bp`Z&qY0i{3Wx31hzdL rAǃ@1iP@y &ܕw=?|g.TR:a)Pҩes(RxJ&TA}Ɣb!P1+IY,(/t&K: m15B B3N PZj%3)'\ZpBF&2$L1M8oC Ҏ}TVNeBbh2 +ěyW$ҡA$.P ID@s kg1TOЇ2M1l<!a3<.T])*3yUNBu{~ЕEprEYŻmeOhZQPy“%IK剱{;RA&(J*hTm,ڸik,Jz]UJ<_5E ̙5jXP5\ Zщk4Ԫa%@+kXFDl @:ZN7h"Uê`C _eKLH^#B 5ޛ ט?%Jɵ1F6-twRzjezç?sy3rI4¸R#Sݩ{'w9SS1:5ա@ TcDvjfvjCPn߈T[idj[i:)wߖ25^jɒҌQIiDi"ތ???㍿(k<&fJƑ)m;ຫNz?\m72`K%8#zO$hG H9tTє2iQ>$sY+!rJntM ?htB"I%OJN $EfViIC95Ri瀉$SǓFV.5VmO73eEyɱ?hHm 3g2HNYJMSpd K-4nÛf:ׄފqK udi\;\q)Eɟ1m-$Ik M9<ăUI* k g"r}]vY 'kϖfk.^πO7+s&27,IM!-.G"Dzc* ];+\F[%a|RP*D|Ps^sh`-SYS+Kc hވsX)|g_}K6%_r!/%!o9o(4O2/,y~QcFmsgnp2wdnf^fd|_-?*cH}ɿhbᅎУHiQ 2l1YřVnxd?|/w&kvjn6߻,/J |p7%RVdKYOqRmVtC_KtkA 4}FxJ_n_4V!OETB-x/{Y5(LkRI]l dr>ü,D98S+Dz&`k wc|[RR#jĈϨQgiZ y,ڈ{ͱ$݀^4 ʠkZ>Y2M_4hC8VxE1Í-ޕ8r+e/9IY$H?C0xN<ד/{=:n]J_ XjMaڰU:N,>+j\–ʝ^~_.l׬p #W~wӅeT? -͚/dw n9iʫyxVL8P(Q? E^_Y腻=`F zHi=-?h>^ wR +cxzq. 0?h5qy-ѕe\ȡUO7v4d(ӭ XF{Q}T>4^Y1qNuZ@bv3ֺGF# 7EְٵȚ쾨RF&-_1:r4Z{DzC2f- bCuܔ=kGN=h[ʺ xu?A46ZjS]+4i5&i=ߔS'gᚊ^B`.4/,sX gfni_u۷߻!k/Zƃ#ocE;xa0VU6X~?:'-3͏8kͿ8tr43i*|+6)Rml'/ӟW4_c3fmQ $x1&i CLU]Ů {nyttnQœC?ϊN&?[Tf3ΟQYУ3k*}VЊmӶܝmBf#k2xdHy +wGbA+s>\b=|0V힊KLJ9/wߊV7 h74aX]xVK5_J,MG/~  ]0t!Zn߅sw6[%G~:Fjrr֍nͷZ_q#`mcyiDUZSu wz]g2䊏X汜жtW+O?˳|Q+K2҅#M ef,ǎ.+hJ7n.aXȶr}%^/|EE?i6i\~"E}EJ臫|5T+. T5ࢷ; `dI#غIi= >uv<. ; Pia:}f}vFqa#-5tb) Vd`U JoO8^30蠔^+cf^xخpZ 9"`E6sfVM\7 S63ݠ x1!A3ʼ4IRH+[&$B)d4BYFr,r[ m,&!5+xb4F#Tm06sRCԎKL䂠h?]Df 8|32( 47\;$H j0sћ(NAZ)$@T1&U^1'`d*TnmVlN nh j OK q@s9)| j$G裖!T9$m|Ka\$V!A,L[줣h)]?Ia0A+ ǩ[e47RzlIZəJZ|i?dM\(E {"ehm6d At\DNU]gI:(e0Ǒ 4)o"G H !/'{Jױ*~ۦur{F^VzVG<[4\%Rt2@FzlJK;EtL{"wh{ Aag`YsꞴ-'st*'%9hZ HF&ItKA錞=| 0KQ-|!߯c"ہȢ ;dQA_6wִ:V R9ۋٻ=q<|ٻ[ 1vߪAU}oRx<r1?|.ږg|8{XvyoYR9%4&U2mx2\6f/za”vb }cwW\co/sC&>a\ZB,]Z}cʖCH;>2>8OvC"Fϣ- Jd`^ͺ!W'a][ixC{huBB޹.SN{RFz+fsG R@Q[-Y\E>\ғWyy=V 2|}:ԜkW8܀eVg(D+a@/` 5QW9R|8w7mJ(i%)Ѐf-f eؒ'(]^艥7 l54#MT-j8fI baB4vӚ { 5=s:QR4t8* +Z7ɬ1hΘ` nSȊl|,Xȼ BF'UnEkz=ϝFoEҳ -8vhf*CQ&i%luD<,ֺޮpCp-X<ɇ>f[Oqr$O^gJ7~\(;p–E7B.XnqQ_<5O߯MJ1z~>kc>[Srs?̟hsVIyp|yHr[ы$W$s+_07}ü̼<39o+x-ݺĻK3"g^ نO?ZߖoaXkŧj?^Ԯ4uզCVM||LHZP|:DoVrdW1۵oOb5Zcx[Ӱ eИs?@r72lş3;p=!._TzX&|{{7)D7͘[.o}7@K-=/ }CHR|Z3MݸYp~q"N8r v" |UCӅT`"T4G7(<o,cˆϚ9p̉4ȿA[qL w=kNyi/'UzvVLص.; ϪF:ZjTsFE,డ甝(@ḹJ_JTSU+{|6 0KnIOZ4$mYgҌ2w,d'όHXp"zzS`^9a5) 1" 1 #TAoOo 9֦Em9q 6=œs!0\DYfm+@4;#tg9ce}7NOIlKOkXXroeXZʱ5gcp%!0erZ+Ã&}C\l{;V(\lnXeCLu@LURl([azdԛ{d҇wi8K4k< 60S(sS|lV7XHV8T>ܦؙE.$/5Ld4de-7Y;k y,Pk-3)V{5;¤׼5DiG Rd˳1`TmŸ9![EɎQxM!&Cs+4[n6 b5Gl\~t.r_A4xMԝ;fs O4 /h+yZ>r1Eku GזqG6hJbVے;xf~:I>(XG tDvt:f084)ޗא?63oنQ"{jP#.`hs-quͬ) *;^C@&`S(UCdSK9::fOƹ Z׬87\; Ѩz/[qvQEg U?-zP{62T]4gs>&Kk'"+EVM\aQI"ę$5Bh;""%I!k(^aUC s$0ٻK|sa5 ҮgimzI\)r4D< o |޾z%d"}] tZ,s!ueèz}M .f_]}z^l ګOd ?&2֋#UyWy!/}޿?}_ήpP?=|'9!DW!W[@q)'?wA4-'ǃq6 Uiڕj?/ eg?yy6of+` g\$%\Ќ0*7oud>dej53@ܖisxwG.eZ>y:2|냗WͨǷ\maL (3/4j䀶S瀶#1uo vj-r)hM3Ok .Ud^|hlkJ lL!}U&rю;07#}/TQLdBp\qxR&% 8Bb (BiT-G}0j. \=toߞJ D}urfОueG=,h&s]^_Wu$.e T-G+2,N6KϧۏOgmpG{W?Wzh-^=yQ8yqA 5~*6`a?M+T3Ct6[BB0Ѽ J1:Zڭ?H/+1"}Jш){|7ҧZ>ձtwϕ"zz6O(0Sf(狝>e NNe/vaՔrO5ayw7{?eMbw;UaFP=x[M`$N:CATOC??8Vҍ^uS(%(\4Ȗl-[Z={bGBK~&.NX==:OF_m8Mhwpc3pk8}l\eNmΒgrv}{hlc5Jǯ^+mޅO`=v80݃67n0 0D#0lPI2(\4֠!wg0Q@%Tj]_ԧ^ݿK C'RM%hڵ\ycM$? dyФ5NU)Eir$, #n|D+ؠgW599rPA Z4-ӴØ)IتBl#7"ô(Z4ץx<#]Hh)ދ=c661%㢖1E ]ܷž=KSH&z!~X$Va篾A9ww?\@;h2V8Aq)m®q-aivuȖrBoq  ->@suu}m* hNBNbHW(+s;ܶf%^|[߸,'_u@w*)є_Lnh8GQj4vUWsˣSk]@7sdc>&͖C>ϧ0hT[-Mpcңyj /^/^j %~\#+c,6j|) NAµ8gW:IPFP sFA΅ls Z:X%]C ,"޼ΐ1̷s(!b$wˡ~iF&cM(ΎgLW/OJՈ=LɄE<?+ h$NΤb"q2ądžU{!9!UC ipSϿr. |.Zڎ[5',u1vc64jGУűz3$_`d?!8sH0Є8^lm(F$$僢J5HmN/ZwOȑÿ"b2ˊ"\;JSZH;M)\Ru@Qa3EI\kSi~6ifF#F/PC>IolXL..}|8w>&Cx"-rx"-E7p3H2JHIh%! P0K/%Vt`|^2 ]ҵ@9lN-//֝n^z puWW|uWWUS>Ed& 4 "(ki-sAphԋxœsX>j;<R yH2FeNEv",at"l0uƵm4JnGcq3ϴ( ` &\.l WT'ÕwEl^>,E^dKxѴAI!h+hb]bL'@> d pLQzX>j;8El MEEt,yn-k1xlHISdxf4NHbT'\QKaeT TW޴Yk7ISv#P JbV k⥲$=58ȔgN+J@3:Itʨs{Zu@t5KOT߱UXh(Q*UYyDNtU6:w6% 7F#9zl"AQ{ȕa[C`LsL8I$pOL$e zhV;pH,/3"J$97xF3$ HTs)ziu\P޸Nذ% Gd̜`ܨ"EH)n%E`%)ꀢ]|rUQdz,J.Z, ZG( h#\kRGf)o @[z|:S=7IdY*a9lSo.b!P0^ љ: ਢ Nҍkf`  MqMu$l3øP#OV}>IWKrn$UY&Zj@f]P#>hMhOr#h ,ɰ}NbJw뤎[k˿яq-Yca'EߚeWYk=|:W/"jgW%SjC=ԛEdef<ȂU4Hvv&m%wAI2SV4<$U`X֥HJR{e1тVGk`APEf2.nL `H濗#[TUϮfrN=Ft8fH`k$ !bZ㸑2ЗcEI|_/p.8OOK._q4{^fQؒz_=X$yv]&BiB8j<`]{3jĠh;8^ D&%㕻 u9\G\uWvkŝ(DU1dќ֟H+\}wv/͊f,-ujmEgt=fXw $ӏ,lv > ,^Y2yy^nҬ#^uBeOv,.5FYȨW:ucz`QJDCk#lBP;kV(c_6 jjgu`Fts;jKh:t-k-wWDwP+=*+1dGh8[7\6ZdA60[zPI 5#b 1dk=ur|κ7jG2;aV//7 n P:Alg.0F]+t஥E 9dy^@9fG2l+YEʀTvϛ)9/)9oJ вP%4 t.*1T*Y2V(dzrv ˺ԩaYʇhF:jAd' ADOk͌klJ$yB8O[KqF`D%-cJb 1a,qU-ܣ^XXD럩ˆ˩ 襑@3rR\kfM.}sbdZ9Y:eV+Ut,+ʘz@+뱑"Mڕs߻f:,D%nV2Eߚ+NA:Y+ɸ -$ϕSpF@K=|$xxUaOf_veYj.g'AR}u'gn4IL~L5H\FeD@vRGud{B042)h%:W >`[1g{ۇ-" HRˇ_TI+:/.6uxnCr賌O>/1DdWYliA3]ŀQ 61\لhdʮr*Jst>1Zж!°v.di^B._'7Ֆ1|C!C+Q4zƴpBКB[;`12fPQ]6"/e%Q]tTG9.<rhf8,@~{\x]]ŃxYDd$"VvEF;zFCr{xw擻su,~S7Mf15,n86מPu};E6 kohix=Ǐr<@9=^GDuGfW:#NJ&+: ,rGuihM|򽙌(ਦ?OlN9@co)]|{ὗ-3l2|o,w4z}+#`lZ6l)1QJRe[q@?^uBR5'v,(ЪEA;FtdL^Ȱs5g<òְˡh'C=EѿɖU?e/pLw^g[i7?"LxҎGʒVQ$DV_9 j7MؘN1K.5sz=3ʯlH=} zygm\cXkY- )"VRu֩#ٝ=U e{m 1Jih򽙌s Λ.)Knj>E:~6൭p-[m<~16!ݣu1>If8iէȗ^o]\t>.TjԴo*A-`!_l;`3MC̔iw֥0*<\}lQ$bٖªW>p^%h!(_mc P!^jy:bLM;wR3^,8\ɁkO S<-W_F1vA@AyZ)rA2:- Kp Ѓ}Zvb(Y?T0H,üvkOeЭt2)`L8;nA_R^j5 fZ%p,28#DpZ$ybDlޙSJ1bRj۵@8ۋC%AD[gjsLЭՕ̖a!뤍-r cNU;!s"%E |'b~E{V> [ն:MV"Oh5#؆) Ǝz}THQGŸULD:f@mw^rNLEsdk=)Sp+ŪzZeN7qM?&2-Bm2)#p{GVG-^A%br;݅ӴlȂ/kΦ0J0#qsMkcwjgg6U< #ʲ5V) x*wZˬY^LtW`YJZ&+ǃ$ 2ĴZ7GzHu*Qz+1N׌LA)oiu/]Y$o/ZALpu; Q–9&2!Hs-TYb p&fQ|넭xSZmJׂͩ-1hH>@W+O-I'ZkY#ˌfhIrn5L+?vv~2NvSdv`q棑0spjbr;,ӶyI?ey}jr0s#<j@@뵱h2O*&#B6-*sv. ${7oa 1JjEb,rƂ-8-"C!l0n:4f[.:<P פM`V6Z:ڊO0=2>mw,A)1HrӃs7mQMU+Rrk B)h7h:`mHa6~#$k5亿 x!?t̠*{Uv |n +)n7}|zPvӓNM_nz:G9]m\Z>}>8מ*ՆO=2>Zvm<686?3Hh̛1ov`<$#cZ>Yq${{J*pH)9C Pd! iAG} d4,޲{G|4^T N1w =pt+u qlopLHX9o/ܯYt[n-tz}>}pu1`|}Yl0sowx뗙❗L/r~YIYٰE/9> 'ht0(8EMqKZ`Lg瞗'>_{>+\5f|K4kMUni7+EPn,1hÅ(%OLeO4UuB%}M,\rIך(_C$U wqjm5U%Ҫ.%!/;o 2EopD TʏgӇ>3?/)]_~2ۉ\ }Y+r5v%ɧm67L&^GG!RmQj4+5)~Ik9o W^ ͖z ZB\s0z-p_߬CA҂V|~Wg/ӟ}3݇p7?;XQIDt-粲,e2& po3%< CC6D@K__t$HTz|e,Jn[XRi]|˻A&Dq2\,g~kyz`GǼ˂Q e,)xt+ʢfVSpl˷O۟N돐%b-߉01Z-Ko||>~f!G3ljx `oK˥= ŗ??ዟ2_R/_VldJ+:Fє_ 1ӇS<]sN .<֩H!Ff =jEi+G:ĹF012Poзj xk˾f6 w½w)97@Qۏ2)Ty ٝ5zy3pPRgygeqR#cRLz_E[gbvim oɡnu+CyA<|jq_ Oj8:h5>gV8YC{'μpq(XmԆ,,BΚګ,hcF(B&k![5:[M8$CN'o IGK|+ʺVC )HDxwwEw'kH;˂?9t )"TОēE_IQzŪ 8, .3PXBYn#R) QWYj ZFaF ƈ.4 Պ cTuuִ­avrdcNgPXՌ62Lzexd]m܅GuD|g84zTdboA2Fvm()LGU;R[V5l"eZQH}ŔiKaq;^<9^Qs5z}z;YԽ-4ImRxo>M[%IG)"\TNf:rV\lDUߖXKS6 U 7pAq\8yGb;/D8zzo爵$=W&5N$c,A\IT*0꡴XvD`|J# }ȿ EO,ڈ}eL4eĈK+/K+L]|_OdrVr]^ Uvֹߪ,z4RRȘ+Snj'wv /&E\b( -8/z'yXU\vȝ[;<&?~vByy F{(j * R}=ϻT"47U(9XiUJ} 42Ͼh֥c4+3z $,bYvK1P%R& -Yfd} V0,H [rDɇz $՚xXV>Tt-8q876`!=`J VԌpa_m,xn9) `)*<=mt'1@q1ԕ]|!HR)j+D0^9ֲn%sOֻxvYqOv>KOq+:q\N\;|-O$Ux#qf_I)!O(@Gv C %:lq &TE<@B\ڂDilԂ<ūTGp h.5t:q{k0])Jtu>B۳Zb*w:7sӄ|P"=:ŃEQ[ L4GA#AիFw^iWU.5h!J,J?>ofJ*tZrGyyNJAop T5~=aM.PD\~Ť _2o<3|rt3Y.3"^ Y p(ų6qO|'wMX\'ՙ I@Xcԁ;Zg4*dz4NFEd(Tw*N+$IUfGjR#Am ",AT\i킡q .zcm$.[*bt"rTK r7$N1,ΙjE0!'8a"alHw%qQ_;m}1%$ـf$ȱ¸-Pi,z0&.$G9%XpG > I 6L闵!GZS " * *US /#ԘvhSC^Hf㙵"GDv8S'@E.ob2ZMd> (AyE6HEJ&Vj": S:ǠroJPv)!:CŠ0T2vE3gz5!:h)9[ $z1@]Y+"^SJ"@R/TKii QSoAV@H 5 &c7)c^_8IP5Z\V@ ے5_/dg4pD 2 (y›q AqQ 820D\m: $ZʥNH-DJ],8 rh(8HKk0T*v]`ic{oH9v&^ݜpjC' F Yus/l'h!' Ĩq;$؀%8;9뮥Z< Ք_f#8'GTB^EB$jண'dC`ZeXBrĒnj\cY@eqE9R-ͬ7  \HL*Q R[kڍTu2KIZ7m;`@MpkPėW!=3NX KIcuNM_ -PԲLo~;v!iY5݇Ov[G=r&P!\kTǫ5/owky]_aΉKُww$ο;[臉}'3XHA;Y؅qӃ3s7qn9N5ܳxpszs:3шP9$W6½ITK߃_aR^0fJ4q7E pvf_bO݁ )LxIo^y.Hق:4?6.%&JIJo~9:sv[{⹘m->8n>bJ_Ũ<n1r^رW_?P]Ak 1 B[ 񼕮`VF*QiܮwyE8 u!\HYPay1cR6iyM4q᝙~9σlH4QJ( j2,m ÜA8ڞJ R1y셽Τs^My7-: w5c=mM1o'st1n2QmN]l $stn2Q޵m,"CKE@>$nEKr騱.%e'nΒвdEuY-w i7ˇ-A7Ǝuˇ J<-iݸٳ|sTv-Fl;͢wsEnEh[+8Lm3kSlQiEyY+$58=! qXyVSl| \2y\©iΆ<7[я<H4ueӔiJ$ *) 5L<Ľg-#$cqa$T^ [uHi36{N ,^Mb_WKSW}J (WeAjC-<,ul?3uKD F)VyBiZ1 U!ȲuCi$ÌRjDӨ\y]b0fRWbq\rjM"VEpVO_8jÛȷԌz`g}ѵKh A gq;y{`2S\\vblƶ O\&S,b6ML ) 綮=n/?$w&{W>!h $]^c%ٵU+v _;2a6cV z.q*XE>u@ X䱝WQp%R*YL,nƢ-EWɢukdQP~3m%,R3XY$˲hK,a0ϗJ1AnϽ@0r vK86gXJ UUbQ*jW'2:KS[xRyuGm4>&j]K6QzbK` 857`ȊW q}۳}㥚u_P6w '4 #qϮ  kX $rkb4ogY ׽zN4-FT&ٲWW"yJrh38j]>ᕫ9mIsAOE޺,u]ޤV6Ҧ<8ěڀfLu ;~7Z~70^^r&H։R<ራ%w`Nr@:ruDIRy35DavE+Mi K+s9]|^ /($B=y[R`Z+{^T/F"ǷKE.xaZW/x'"ںci-!M[|Z xcC^ +yEg}h&jkSO)Юx9@;(Q^:Q'J$wpEO;"ԇy@7/  ^2 ,QFѳhè]Vgfn?h6?MDU7*_Qo]' ) @$5a^c}2ed1 AXhH-VAe  VkL҄ Sv^i׷pml~Nq)eR)Rp%LWz7!p怙5 ;j0[5QH ?Nd[;aHxgOg`*%xnpF/zQ0C3_ nSLu'^^7tqtcӗ?zI ?{~/t|{??OxǷ767^d] tUϿ{Ջ]?ϓ礫SP_jҙDLp>K1H;O<踟_xI#s, g_F7`Gqw?3JGMۻq>N)|O@D&YG2tٰ~S!QDW[j9q|rT,V8%|8=|ʝ{k~o+ QɲI{aýl03CvO`xm9,?=;^Ax[,Vf_, fzt>]JlqTtbvL6\`K rA. 7li"bHKR~wԤ `}V Ŕkv[T#'l?Ci J?-(207 2 cP| ,<&,ef#+ťS.^]j,, C⅀&mĶcܶH ˃-&xXjI>cz(P+j=hʥ!8 1@H`% Gfًi#*a#C ^y/gGF=#ֿ]wo֪VӽҲߨap_5uJLCum;؏p"&{i|Z onʑ;>J/v;R3("s>ҘqiP(nkTZ*+= J] 쨀 "dsZ7.=aTZ=P 0ղdv|dR&eu8V mҹE4P!novS)zg)KDL.T.9 p 2\.Ƃ-Wևږ+az|@ [ EUv PQ!Gs?2IXRM;asj&)9#YA@=Fë9.Qk#kaġ \Yrhj5QmRJ.xS`#H0Pm87ߝ{ߟuP[Z3"S?``c %25њ4R0Y&p7I"o:ghtJ>;w7 <jO >Zs=x8 Q+abr4NXF .uHnbdWր 9S5z9,b :7WAbGǭ_zAuTZe5q[7x3T*LJKJK6%chxRJɹ*ut*TJk111퐍1.aȟfpXZ(t=8ٲՊ )/2ljEamM<(4 * Ή N1!u*Nڿ6,oU6tX7"P^ &1 Rm Ye.l n)7sc+Ŭg1cx=*b{Ϛ>@tC+M^OTgg;mTr[|3cEשWtB N?L8%#Qgbj+tآ<;ENt؏F6?l> An~vYFPHT]`0VǃEެr Ej0 Q%}`"dbBJrt0vy+Ŗ+\~\5SƓX, Q4,C¯+ i -V{*r`D>Fa+ 淍!%9{U09ͮZ5֬؆ ߸US=vإcN]fXj*&-b$<"m"iBxo aadk$ U>TժX)̪Xd)Wg`qМu5%&[j!{ݸδWԧ%%@(RQ25 5ž"|IePLdµ,'e &XnkVuup!|P8 ޕ6c.%Ш;鹅n<͠L:ݶZ/U%HYN4UR< v&tQ,ULSHa̒!a@z vQo/c-/aM|~jו̨yu>Eˢo2Թ֭no[`>!L mVsk̩|s-S]y-B{6j^:ݯH):y}>ES~MʙzdDA!!cĸhb#kqL4Fٍ~~q_jdWYk"#LǏWUn=ueyݾ&6wo,<3=KOْ3/ݺRh?ܖ~cMz<⏬ 6gu* e7ݒqp˾[Ӽ;]t=k-Nl}1"{=/ry4Cm5ԕ=ݩB/3ZLߨcx:-nj|̐7>E)DOM* u9Xwgun>feߖ|:q>{'WB X*Wz3.v% ^ݹU]C@=[-bmw LA۝jo?o7=D8"nR$Tb^+NN9L4I&$(PcŒ,1FHq9M);U!Ų^kІCͥGܢ~o8fIPĘ' M2l$¥6N9JeH51 瓻[L J.9rl?\?źyI*;ξk Yx?..[Zm>_](7J!;I3RP`#@DkT adz.}J؟aB5fI0R.+o].:? a10 TEm@En֦at:#3y:G9zE>&dQGD/ćo3Ep|yep»\ dw) a-^%R54*0*8V;TR=|sEpLK$(_*Gt} J~km<>^Z& NIN(t >]疚8v&zm]uH 8:qND#!$)4ipCq(%q9)XPA#& +IT `t5z?fo5ER:>HsvZbRiFn} ~mbkj-zQ;׋0jP<躼`u_ڪhÇ:8%LF|Lԙ?XlҙN+tdmrǿsO?ٙϞo\k ry{ԏDov㯿hWfEu홆P 4j@VJ))Ap:JϱVpK? DIS.dXq^Đ:;Yrv&CΦ6y[`'Ґf:It>=ezO/i @1f}E8'mȵ`36`rhhKQ"qsW RZg{R`TB)&H rm&&|}awX?"TC8‰ ,x:>~1lBO0C*MIӹ^C< {/A)ƛ, S<9W15$H)-l%,0ּi*ټȘ {@3YКf`dY*%ҺˊLXH>ditp)N.-NM/ .M pkTQR %M1,, #hII(")*VEWP+CoX)AvIZ7 Itga8_g)a4`!3Qך);(ՊK|Fa|rM[nkbʋ]wxq}sB]<4Q rTNNvXIŘ)SNjzsZ-߸As\).K,G?d]'U8ObVDnzF);WQvJƳy4t SȞ=֨qH7.K11TX#mPClWsON dFl! u#t02pWq }ל!M!SK9H ߺE%/C*kmȻ+Q$K]  Y!ܢw JrOi5'FGHVUm!3AdFNB~B"j 01\߹O9C}ZFvZCNq^n E]}7((,i`)Pι {9TVaGeܢAH$"fwk\2- cIZʢR=',gM\pE߸@ȓ%T}@}шbJ$*2sHMx(:@(~RƸc]QDE0E+ !OLT?miAмUb7) ?k?qϨ0@< ?.`m!nZmMVVPS@ЂdN|Jlݚ3E'MwA)@fNA?"!RZ3v /yRnKH!".`r?:ʴjgSHKaN;gzF_;aRS OI1YtЗYuձ$;S/%YHJN%uSuJFG C+DP8%Td:y.kƸEmE[asb!ʥ\t FS*r \LW@ssZ)UXɔmvMH_|*bDrhKsZmŵ qh|gty/fe!x*qcI?4!Ttzn>Ӕ":oFT)6#: ( 8)lѾ)nkh.y7qAQIʻi(%mZ0Ow9ŤIRH!K!-d|lhDVbz\v&}6c[5d WZW_kh9[RР0BW8!PU$>iSe[~ X ךe McUt:UaJ;>r_C$l;A$FIoΎ蜭VrɓGMjJt2RfSe4*oaE!t_q9"gyE+B ]w?Bm4rLޡ:`M(]5xY9q(a셟cjuCvS4Mхp"QjrdBUƸv论\Fę5c>cc6 z)1`'_izߡކrZvXQv$oAӱF(Y.d֭؏t)&/ANQPGh|¿N :QuO'UxV״~q7dN,TpSMNFtl]k;jfH]\N|en(λB',2b(7KT =zmAVJ{JsG4xJHɂ҃5uhWnGvxKDR ʎPϤmqj/?Spi{QHƥ{ T$!\!x҅NBx܎+b˳y{6 6qd"ϯU~J|(v2~啶[W/-f<([!Z"f$z . n<2cx ~d;O3IJ`wbgZ "r^NTf)2E~zB'ǽERZ[yJ2˛AEqx.X{T?;ۗWC1jxsrI <2' D9Zl߭mD Tk BA]kB$g#J!r(v R GM("\3IRLY!:1i=?^-kfwRۢD y!q&xÎX!R wɏe11*RbRU -a68略qLZ#딏A7W!$cJ 5A ~vhuXvc\:h%NTtJj)x5xyi8zf*g̟0^ !rt/1/7wQ3=6G_28:eY5D>Ѿ?ZCv[C65{-zT{E_ΠZMBwb&ĹIGqף:Czy71 [ayjs f5 (pe;=9ܐyFyjy=1OK*e%Di^+gbj~ZtݸG).qal#:.q]T\A_JU o^JFpR0Đ;$'ڠI`*pFuH,0Az& ol8 3__n;N^.[c :%#}S^W="vUDhrXeO~҈!lnUs_)~V44w&m_i-HJ(UҬj`Y|Sq;G}Cm^M.‡_ 'Әo XX;jsCFN{ ~#>TBk[ >b :HׄFDZs:%ϏRG* K"7T i8w[DFh"UL$ U:B [< yKOKO.rzrӓ\ӓ 4 rQTyC)夕;Y"-FKPI0 J[p*޵5qˉS]\%%)WNRyrJ5D{w&)rD!W35F7!OoZqFKpJK8摣7F]]'A!t9O*:+HèCleo1 b6}ͼ7 +2B7ɬBJ!S̺_9#~:j6?lJ[7V#z^nn}2_tJaK ݣܨ}z+69<lM߮ӡVS{?EJHȅhN ڍI(5R Jtj~vkw8jpkn $ELۓE~G)RJyP Ft~v8e5V{-\-Ss,PckM{? H5 w>r6=U6kμQv/͕>m n{ ֊f(|ϐ Zyr=\3DrUuΗ:\7Z oG=[k5Tu*0J%ZW1Mb[Zx O[Xs,IIAkwAS6[M*&W@ g7{jt%}Ԫm(8mᐒ+8EرBчU=2y7߭g1*IiwnLCy`0J[p5ޯ9_ !i4=}\=,Pb'?|v{?ѷO߆4˨Ko?y@|&=c~ g!( cPal]CUyH<4 Mp J/7`Tq X: pn[bn4~ q37_J6~i1d "2i0ե3TKRH&RS˜ g p@^ .%(Ub89!κ MRS帠X@<8+r!q "[i 2M:ȩa2BXHjpƅ3{ă>aGjpV@@Cqw.lwΎV b-J ,WHEU5ГIC5S7Z`%.%"˙ D(%]0sW0ddͥ BHsL MqT!('ZxLoM\ i~&ϥPJ. PJ.+an,+$GXe&܄ZRT(i'؟aVk+7?(ET8C^Ӝro+! zU9#=ͦ}3- Mrg3ywjX?fMY,^1Ƚ;G9dPΔP,}y3&p1:!9#Dϔ_~̵{c]{K+.1ĆnzLhVxѓRp<$6YW-G*k(=kDviz# Z(睒qNy~):K#ũ}qcn7r K:l7s #c l K "*(u62\MQ;#[9\jowo( ^+-F8{G|nJ*ݚU/>Yn?3Bs[ΛU7Es5SbKX<XזTXc,&mH# I^_M y{]mx lҞw*BB 5DK>Fik-rwN!8!rB\J!9`Jxngmf}55x+8Su+~Xd_f}E1V|pp߸SCEʌRlf~yabFK[NMZl،&'*a'.Kvdr>C(4\N,U5}iۥ=}zLK59h#;Vٝ=Tɒ3lw+^_z%W~bĭJ(*zFϋ8bl(ItQӭUaTW4詃AV.S`Z~8ո~>H΃3c+DTQT ײjW(>z5=$AvxI<%!k5C=ѮѲm8͟ԯI` Q)K'rjUX; 5DnGU(~e3^^cb)(V_2~% ^MTt WPObrb~Ɖ{Nqs^Րv{f*fR#)>%{㽄VuD!7RN |>"0IG;YĴT;G7 9 e{v̍{k{}PGAyz(kHщ5['p0Or &@Ě6;IQ ѧfUĚ8(&qQb(IbM{5XE!A7h(upϣSg mSD *<+7u67[5[T&U„$[3Wit$$ieIqMN-5I@9HP3 9HyiCTM @CAQC)6P+NT(|iRl[ aP_5HJq^^5wIUf@D&j)rEr2iv,f&59`!ÕjA m),a@ f TpM:t/$> Ω-,W9ˣ?Suv1]"}!bo:=d>% ~SZd0T6ǿ$q&E=DƸ/Q_VBAf_uG_=})sA9+vTV{ TUI^$+WTDR9g? E+?LhdYzҝ$EuB{gQ?'S\r3qMhC9{OC+S|~oʐĀC3;$ ϦΘA3lp;we=dyû{1=olɴ']u$Au&M/;_n=%E-7inm6 - 4cRۗZ<C{\lg4? q5E΁o| T,}Q;`NdXBNH$"e]!S"Q{E̱BfHYd(Z6k Iw3#7%ɞDiJ#I{'A!D GV#wHom_֥1ky_TH4x$.ϘS^6FCZ- aMmKOuV0ڢ{f(Z7G0; NYRXeUK\Uxr\EdClWvHKjwINB;/<=wur_**!0Yq@NzzWWtГ; D짍!-ؖ~uF]  0%o~t_NPQcxbܯR~mBic'k\2T5q42vjY9!kZG k1 F'k r30oWSY=><.n_xq[Ā?q*' 1I}(*c+"1H[끬* +k|PAHz`R^x/QWTuYG^P aޟKK Ӧ\ǪT޷ft:T[ɨ*`B[\s'&2Q@`ڂ{fөHRAU H2:Mf@DiMv#)@Ӌy"tocxg. BEr=G8@FGדiTܠI Pnj[nB%2ekT)T()V}[-qTk|J0} &՞Q d 0\*7=ccp8PVn` +fLJe,*SE )Eϛ!1,1ػ.*"7W>OqeGf|ԲBZ|ca$a+Ք4o./P(%RvL 6Fzh+,;N@ڬ[-f0h`c°HBkYd`u,sʄ꬛gqz1K|V3}0Z㒈;>@b$įu d\@1LWfJGmJ-Gu[: >LEKę~0L[Ja#`MUѼI6pzoF#I~a@Bח.(%=֊f+٢v~4>x4.A<y> JrQDܶC خi '(w./;KD] ;aV1'"Хx>!ǡ?H _[mBPK߫s- FtOZ@X_]6']3I`gJ WX +[SBke 㾹 B!q1&:a/P$soDPtk y)9۪)fXR\2ݹc~%0 O˽~}zMgOX?8r^V "dL*iXF4šCJbs>oýۘ7ܮ]gSxE^؜"N!&3U)Ms4㓏yá,녜?yu7~ϻo8|/;jqV1($~_>?p&(տ}-*L6Wۈ#O$ .O'F/4dѫܸdj34k˰A)%Uű#~9lI~ o&-o"aőЪȑ" Z(2lDx5Sh i4!yn=1KKoo߼)a\L'nP_}x;b^>vTQ{o v?_3t&o96NPݒ0 JfBWJL"u*MK%=BH"S(G* +Jvuְ)&9V 0H'2 5f1Pn:&f<>0\F8MG&c9\r6r?/vs>i p F@Ǟ3P7Tq:?7Gx ͍O/_|0lv9-?h $oqW%Zkl-%[>k T{[\}x_xZ~ pZAV9| D_WANDy}W7ܽ|Hoabomoo_7^~j^?7o\W>({{׷ot?_>1x.slR,i? q}$L4Edum?M?O|7ҥ9S {s*֌N~ys_ 2qާi#G=mNjpXa~.ɔ9' kr~6&*2<4Jj̤O'ѫvN]?~b0\f`hj7N|aڃLG|vmЫ߾_0v~<n/oּۏ0OއAͫvwmȓdvdGo{ݷޠc!'+ Lϯ:+ndx~i*{}0@v7}M@CEWUЋ9'#N?=OuN]{,,#4qYڪiʾ7/wW|vtsT.mh33n6],`ߟ&2灣.+LaEKpP漑|nmdMg)?>N4&MgZ,>J0?7e1?@mYGEk^>|,5pqMaV1^Q?VQ됢~PU%;RΉњha9#"~p’NI9Q Y&N^n,/gESI8-v7n~6}a?8LrZ/煷wS@c7\/= P>[D.s|ϓsATbGcED3|g/sL׃6w\yn&z\b[nԤ>dm_ٽ AmY 2D2]YPY҄&S1fpV)&RdK_U*~*EA>bSmQ6T]%DqNX"{_et4 4.vBV‘0<urנc[ux: DIDŽb0tzN;(Oy2]eeLZQM}^+}e3\WFbҳ< lFe_W[@PWTS\՝ll#rݜ`F(˰VJ8{2pȠRiFeJ> k2H*D*D͎@ٳ듘@8? 5'DAF$PŃ[{@Bʲ .̕Jqwrt&!"Ny*B.:y$kFA̩k a;?>bmvA33r|z0{=XZ0U$ETDV*x` UKEH*, g")jP" {\v5* 4̙(Bi04AX=Z n?/|c:2y)mGʪ'Lj0|iVԬxv^9AۅX8xr #dȒDŽc_~KϘTQDs?``׼"km_;,k;X@xekG~ ,h\TE5VX>U#,r?1S!xٙk3f3쌨֞:xTm-nbA Eё赥biz]50Foaa︓x}=K57I)9loX݉9E80פ%&NLtJju T٘^9S" :;DŽ4EwY!%[B\"c%99 NG{2t'}N%~/ymRS vJܙ1ΎP=?}X;+22 h_unFaы1kGaN 0=[β"mrjs[}buRr-3xI8f )ͼP:Vՙ$k-!uW-OtKhQHx(4qL41O\V-'e&ruKu֭:֓ɺzPu֭:Vuk1Hz`[^aZ{C)'n4tkE2Etne.疅Z v#|=_kDKpU~㐄La* _D@hv:J :*x8hSG.0hW|!=vSAyK}|&cmAnW d61rsUx{Lxa$&i|ǬBܭ,quތ<%;aeLY<7*WK8 1+ %:|trwV2ow+:e^Vz.(cUa۔y%ۊQfqdy*Z31j Vm8..| 3[:zĦ-7 ?TJE.ҋ,Q(=Q?$`By–HHQWfI4H'fQ#h9@iĵ)  ( .s5DFI*p 5nO]'!x[x|w[#n~dhCKW*!>V%y%=~68b{ws?@K#&/#/90yAȃQ##opNŇ sٵ͕= sBKpKȟy %yuAh$J1{ӞAʅcaJ(!;oJฤļՄ+m@s=)ʚ^<%0d#Q^ʙ0h /[܆7]1]lX79h_#WCb SOדL :1։x3'"C8w>FCwc5 6Of.z(Z.34"w$>LuF7AM}ۯKWL!r+{`7 I\[梏 Tq&wi[V>0+%UsܺnwR)yNJA]2[ljhܑA&"⸈yZE>#t5`6ԽJPe5*+FV89Qb![%1'&nҍ@RqѬ4oqy|X'N{O5Vr? rDh̕רukR *khjj[>*WLr54em-C]ѲV:,zF_b\MbzlY6vPBpcN큭Cѯmݸ4؎qDVAf[Њe,/ c[_?FËQ֊8Y>iWIYd9%9޽*fI𢔤O[@`#x-p|(2|y}-0"K^(GN݅ǔM~>L^^f`zwmHw^Ǔ 636bI`z#KJKN1b,V vM*.ʴ&:cHK25.j&^:OMZa6Uc4XRr٧pYWց7i<6;3eIJ 4S++ENSH+iɗ A"nhc!O}ٓ2yQʍT,O[Tv,fd꛼Zުy%TC9y}}y|D.OsTsNm*VcϤ g0-Ukݫ6eB̑9_3Z@?&CG·͉ki;q:0.0=hˬLN7g16霓gs>γCVqim(Ho|f:8StZ>6>TV6z}_FEǫ%wȌ:pCîSٸ+Y*O~KpI@.'wH^`xV,Ԯ?]_׿TxfK\E|rSt=4_˝􋳴Y_3ue.Fޝdf6z툘KDC$hVfpivkw g~N$x';0AunzЃD2@^Ǥtv:HJ]79֋ërMm6RV+d{4Pc]k{ٰGZNm] U^8Xp}jSv߇ov_OvP˻e3ӛקɓ׿?|KIeMaHHTrW#V̅}306͋鳫=+u_᫏Л", y 3zwHCkQ2t'nx9juVto /z]/וNy}ir (1{^q/}>St:>5>?9}|T9KGǥnp*ntFu~7b0-#tRy9&reG+]*0td^Ӊ=\`Qa &<'*:Ðo'[7vk0֗72 S띠_ڸqF+2ۭ|0!Vɜ MYL<}Xrzt~prU˒Git.Xړ+KQ윫g]o#@ˑimSE*+DFg1ɨQtF]MRrݕp"dr@{.&«Ȧ$u&v5/0üu޼iSxRu to1"r5W6֔ yz==G/-t]ud<5[&Z 3V󦾯l73^aZx?O I@s/ɽ[#ן"@q:=xv+| Qz^^myծJur9Pz2*O*)aH) ) S$ScBt7HG2À 4Fqn}$$N _ޡJ|J`AkD Z c#a9DcPp*hHx4(QtCyMqKxf'~gPB !Q xT%`.Y%"x)e҂b`B)mH3lfϤҎ|2y/a)HY?2%Ye}a5'{a#1n%L^xJ_,M״mwBq"fhub^=x:['QK$dﻃ_%%ţ8k˚P] I6oaKȱna 6!qiQZLofw]GrϬ:UR铫M~ՍM4inw,2`2n|rr:T'G}i|7|5)6EhVbt}o"N~׷vr.psN'I kHXjz^91~rv*aN|:+!VBzEiieuQiM~mk~0x4¿gŰÛaNšY" 8 bQBo΍`YwO-2-YNZ[2Ux<౬a;'4OHɳVc巋ohh˷3v׊yV W3V5=<7(,&fEsF[f"6yuTMkXkqM:wކuN FBEwVQC A~ lߎ^='&roV>zlgJP$v8Z/vg{YtiQIAf+ 2dJǩD2*wOBqR&'kÚ5 Mv8 wrb,frzZ 55 NXf:chO_J$ ;C44=|'(S>eFSY8Fq AFRB̸!4 c AYA  6ĸԟʡlj܏=Yӆ-bb ((wߨl5ꄩFK&۷u+VurV=g;H(Dκ|;r襴 ^N 42į}Pl9$!p55~}hB|J4,Ftj`5~Z8q<"yɐY*Ls@SU} $jaӆæ/ J/N}oxn3$()2vI1B bA r)@-rFp} F,8%UNKw D9<~hVwĂYFh50;f4G@/ ܹn@,EJ*4 |[{@x|QbPeV1qDR%(|l4ԃn%!nϳ6>z۞w I.3c̘23L(D:Y'A'F8:Z[ÕMs3)IH FcgLjd ADaiƤw6j101)u(w RyJH4'eTh@S%!'CA16xP`v $3 Q' k҂ybjE$>Rk"@b)ǧ xy{0(u\"KqEl:5w˒f-vaC4%; ^FGop0 GmHhlfa~X0M#80 e&-3˽}df`0&MI<63`vxn#W[BGܘҤr0l Jn̍u&`&CM\$q2ʴ^#@ L3@CC$ D*ANDg`&ӫelM3af~r$RQj h4#{+-sj&g NN7̅_pZp 8mGGMHWҖgMUQnS2T2I$hʊa$c\j8Ô9eFKO[P^Rw'!~7h &Q`8,fxZ,0W +prk@63&ImCSY5HdUa+z#|%cYñpldD-"(0{A@9m < a yQ5*ϕ͵@=b=?%l@N+@SL 'R#\>^,}9ؚckyJB.MPZVQ,L-(HT%L_UYrhTr,8Jb[ p\i,,7% e{d%^;9I5ˤ정R*%,)`|Dea3z$qW;̓3`T1j X.b+(H-+plJ᭤ o7NY0aa` Cq`ENE΂l*p0BC8`~h+41?è"!(@8Ԭ.j0T_x2SizW0( `)Gj`A"XLa T|4"iRݸ,gmdp'VifbêFf_HׯR|:F]}wi'V-od8(A}D??==x0VM1.:6SxQ)251IOϒDkNk& 4~' 1Y4ym}Z"bT-sxՆܵekpcJ}/r5]K;D1bP;WD$ˆs:GW~ıeN9R,Cz4xDreSǟ+#h!uQVN;á#*;uK;^!@Qy_9#)aqұvi:FX}1CPNˎوJ*xђiB Z 1P&4:1[q\ݓjl7nW7. sv< i#}aΖVWM= H}ecxmXƦҊXB )B#f^d^ҟ^ "`A@ ,M; GN./(]O_Kp<{ƞT90''֜)o,)60x ΃s#,.έ`6Xt]wՉk$%-~ <6ґ;2!V3J*2M[TBDP/q#ou1(!8jE4S`^ã5myo-܈Pdޤ?`c.[DCK;9bJ,79QErVsHB ka4F0gشcD*2\HhoY2ɖr[pd>Ap8K$0h܆OpBt; ֟뉚RL"8=='k F) GA4EFb`Lp"v`(J0eۛ/8Y)~im[?~hRXqpnAdF1qRr;{F⒌Lo<^IPpd34S d[:K%#66]$ś?"QX@cf=yguɒ-gz$#M[˗*5˺_|c߀_b?кfug/W/=ϋ_ &yM)5_O _ xmofYJFôc:dΗJc~sFA3K ݂/,}1{|rgw~zU@ր zb˛_~{tmBuVnݪ'*CT6pġc"ɎzF8\)r`ǭ=a NR~Z TS|xw|/5IQnu'3$[N[nhlWoswL1ōtš80S4<ס$VSBs/[KCS\*5bw-pw~X ?@"Mt"_u)>,_nŹ<}z gQ}J.]KȏE‘~ptxozT4س:yQQO[11f\\l'Gר QUo}ۢ-C3XYc,1ܳsr=<f7y6+ n7ꚢPض .(>Uk \}(vRV:Q+b\dbG_FU>O1?Xf%.Mw-)Ab`BX o~6hѸ o+7YfBnCk4l 8G|k$Kɭz{эq1DHL翺}#C䊍ۨ `mp깲Y2y0R ϕb_ԝ-q R鼵J԰9k(ETzf-lwǩ\h%nM+F'#ME.7??ς[Յ]3"GUy 5wa5]ckJ/[h;{U1? q~yQ *!=cW7/5E˷wϮ1(sKLfl: 8MvȎ9ONrL49(jvwVZg!Vd]2hZ"w ^!dkӫҲ`,e#6|Jb/)%`N ;|ou%ʄ_7+Q& ĘQ\*= )CRf%usȒy\ƒl!jK]ڴЊKtlmLlf6kv'Tzz6EvSbz4NIߋ$6uF\t"Iyt|cjNjAtȓI7&O4r"}vd*R~84Kv?};EZN'QR'e)>=7I2}(\bA?AƔ!OHq$Ju"M1pmg(LњULTyw=%D$(72hk +LLD&a{vzV9n>Ost"ip=8DbT{[{mMmBv(9JPS\ϽUS[,뱗MjXN {`S<գv2j׋T9z,ɭ:|O!48,8KPX!to:ڇ[\"QN?~[/\LޚIScM#lrI2HRf[+i %,>Te4*O&IV l' *18z `2GMv2(O(FHʦiĿ;Sk~|ڃ[#&X4Qg?~(688ESGGQ{%t>nਫ?v J8tQZxLYEt[?^+MeU88/5&ƌz ^.]쥗I dqW0Qzs1!gSx 0&R?O٧Ţ Oq L!Io^fLq㥹l4OBgĭZlak1(0*&JVbv)dFCbpp;mvLRJ;8dH8DVEf&{B$0 Z8^i>6 @"Gc,  nG[R7{u9ØOa)tZ>BIs>#*\N&s[&5S$pPĈ{6pe5,h&40Us9BZ k>U5)Bp^!bmJ(:Da8<\PYovRjehE蒎bsʹd}7~cF+0cT#9Rw[%%Ŭ$VsNJ:VwD2  {03}[9(129*j(B95nN[ٌH6U{[,c|@ZbKG->0TykPp`)^I91YI8ɑd@n I^]Ty;~ P(W ˍC(`DJeփ_f,p|d 3ƫtG=c>_MJ䯅R~(6.F ȏc8q)PN:eP }xmq7& 9"BLɑ$~WΑ>QFcz9_!2;;SRއZhɋlp݆Ґ=F(3ɪ1dV_ƕ5ҋx6vXb(]v?6Xvߙma g7RڞxO88/js;OwXsk|}-'#[Z p?ǩP6^Ѱ5唘DF:TE S?ޘ+۳ݸzE"YG8n֩\'8oxϗqy.~xS;oR;?$BTޙ>iy?w5Ō8K[Jc9C4idx 독 QF?)gap+b+#RK ʻnл/˘|1)H"Ն:o"1Ud_z>?Cu.^-SwǑ=; 1F}'9Ө3 I ړ pp&.4^fyFٱV!*k-w}mQ0͛ s:A]MԷ|^NLwttlNOo..6Fr*^%5AwWI23+{bn5[㲒o;L^݅AFi _$@d?cupػ? )#pb1Syt=D6T)%/XQ%Ż r $ Ubn&"kpHEW1}{zPTnP8V{9iy( 樟`Dقmj$~lop+)K1f-h^ƤAh4fd7R_\hp)[5nREI=_H)z#))* 6'sC:D5P?jw{7ryW/{k ƃAvJ^/ў2 ֧0-Ek)F\̏-A/~M5ǃ[7CE.k8ځ겶r[Kȼoq\eJJ I)@"$0f!2qøH6jL ЛA 0:^ K<ԛ4&keڔYf{Frs6,X*oc,t`]iR}rYhTր,WօhA3^Ӗ4kAo YVe{fxp[*jmTY{\?IVKKyvgn/j3gK);ωwi0Ro(ګ| O 9Z 3Lq|?*yܖ-,^G&VQNxΰJRռ9&ƭ8"!UӜ>\**3Vi/RVjq˙ (_pȻ;ho{nVJIJw]#B[Gx] r4E6yF"F- CwB Eɹ+i{vԗVT]J4ϧ؆kَZ'YݰsqCO7Z hTJ{"Uh֢{% Aۧ hQ+NW,&[LZР9%!^E"[GM*nx1ghD9̌̂"J VI) T."|Y|oM,3i#l0f6WJ2c7.Tkk⒥㲇Zc YB2u̿Aڠ=+kT15Kdy@,8g,U-Ô; Jrr~wXh^xe!b3uI]p%n_[V S)ea>"6^3NSm)j+ Ejk/!7m}RLғuS!B Lv(0pwpQ`\Ӥ. 3C 3ڵS&Zp WAa-&~ӝkHraB DDŽ<4Ff0;+Zc*V>Bw z'%ɊbXkUP9 FmaĻBv1p$KrXJv!pU=2w5l=W)i-y9HhEjC92`HHZ8Iii5?~| O^ZzLh鿼!R_uu{oݳ.l8EF NV0DEg2_<% je0L8DDC4:%J#ND!0=|cb3]F#w| pzK%֝C3cL݆JmΒл0(Ф`kmE TXA,leo,RAYܩ1Q$P+ c+0.cQR~+)*LJԡ^&uJ(\; '8>~%2JF|I)?$/w9+)Ɗ1,0g?Lxp%u3pPW Gc ~|g}*HUݹ; 8<'qJ![3}7V O Ls Ts?fO> 5kVk19=/fn~3J#P |}IILޙ? >xDTfBR e*u4zѫ{|^NfӨEbȚhz"uS5r[u78*aTA7fT<NO04<1aU(EZd:\04c8Ҍd8% HDa1Q!hڠ1l@ ;TGV)?Ww;8RA烵aayO" 95 muנ|7ԠO 1~ +CxotSDƴPK ʻ]+6Ť*`D("ưrjC 5Vx!I C!_gJْ5'sz&$_2Lyruk͓yMVK&J0%V`ݑ\:݌1 ֚_xڲRwUkc2 l}8[O$맑ցv} #D| KLR(zsp#C5U\\~wbv&_h|7N(EOdXy/3 '^DysӋ[^N,1b425[ioË0@X;ݴOGc&u,eh: H+(m RV^zgz`?w5gMg37v~L% u÷4i5'caeh'rvѮ1U:FGp}qBAToCX|1g#E1" G_z> qֹ @#^p~u>wb}`l(gQ9ggt*~n:=_{wT)Ѽz>D8m0i=n?'2~{r_JzHc(d."j!A+-Xy=}2۞=O}i㹻5ϔF />4csFh`\2|>˹: A D~B2pTE-EEYI\ 2x`J!W*¾|IL5b5{BIi bRn;@8< B|(E5>P^Mo{ϾWk/ۨH0DL ’@FOF€BXQ1j} \˹]]c7;;uB4(-B-_~{S*RƢ)&LHA2"AFZ0##E!BYXXbWrDF2S'WΙeɝvsY޴ck]YG((,vmmK< 9]͜9$ο$ }¥rIckWOh͝3JN`1F|ڀJ ,ʹH;&Z'>D]Uwa7A܉VDc's3 z)NWx69GRZx8]J.w" KMzEAV- Mܬ l7.܁ 77LW$~~2uUλ_2+)S޲?p)C7 ea`LOax;:&J۹=<+1/F>: }NB{[<\a^to. Fl:wjVJELaF֞E#9m0qMRdFN)JZ˂@/dh2ԷT%Bh(IX(FnhTްdO"".9GL Fs5FKxBBEX&kOlbUXJJ(bNYEe۲RߙY9;塰}=UmOƌ 7&N)A2q,Mdt3fEԱaC$Na9)AB @O+;}J)pK ,%YRgK]Ďi%c2)b56N\jAk;VR8Y82f"+D$Ya."5V΋G~="g(Rh"gV0 F5Vi B{nBdsA sakNqKWl;+}P|kRg/V%!Way xga}UcJZ:{GQAmSn9A:װ"ڙ8z"[ A930N:KsYSبXvX -aM"F xENɉXJ""Gb^PC[ƽ!2C[rWhkAn.,dDf:TU1 A S;Y(qƘT01q08$z%ƙVJ ɼ3V8jw7N *1˛Hmސfr3cBMSc7V޵5+鿢l /C:jR3)R$JdJ(粵}ɖ)@"bDچhEFȆH:"^Z7…6DWZS!6ڄH6s?Ir-jy8m'Fa.s^} 6*,TD/&mʀk8&]Cޗ@FC u`烅Q% R{6[Qv+l8:oa(2B#(Th()G#*853Xd5f;~QlX $68&PARx@=U& s7~-L&r6>A[pOȱ):y!fu=)qu;M?1:')'%fK*8b)rHM# A,D$Bp#( eWY6zҚ-5+:F0/Evr2'/'1߳׹}z5p2lsW*؁>JM}",U"fD)Ir`Ȯ !P€85$(F9׮T"PyE OEkA9nH# s`?eQZ0I "|ɱPIc -dτH=Խ^6ؑ<1,uW<"AD@a m a}k RDJ\&ۇDxse~Z[.' /]mptci3{^҈i+F! C#Zfǔ>vFq0v3ԊjAqDu8@20`[KP(mf5dr40%ZkƠ ʌ^|g5Ǯ(-dž#HiÐ0P(bZiNx[Op6= p `~/_+6f]Al?V޷i=`=$l= aDyT H\G|v.8iǂe3P;~,y[ ^kQ&-}r"zm:uxO[KԯWDI,tX˞gVA0uK1kYB{iLKK/[  V 0וYdxhn^}p #'>TX⏒iRh/mͷܶoWhkBcD&4sBs¿c'3|q _*az{3 \"xlYk7 ӷoI>r䏅]8W03jujO~$R q=z?Y܌E`\2Z跱W8YN"HB%-NJ,uж lZu؍,u]qHkT(.PzUڬ* Ms襱L}[z[[j0H*rVQpX!s$&W,&^÷iŤALo>۩> 1%{oUE~ߚ@iCN_C>x:*5:DL|3wO'}o"{&&qzO I-˷pdéǦc-n.+3ͭL|}"4YJ`}5ےr&rMnM.I"Py6t:11t\̾ȱ PƀG?nQ~n?Qic-Ddzs4aͲ~Y({N=| Y=c<㙨 OSfzz! ;^7۹ xikX;i @ri5%"WݦVAu(a +g~',3Q* GkO'p5?;ʴY҆AO]"`Hy3M7sݓ22.b"Ud*K1Xv:ςY!}.Ll- .Ck J3e/׷4w/ (y{(e;vKbBa )DrBJxӑXӔUc="dKpVF%sBy< EIo*G=xjdGN]3ȋ-p`}J`9^K؞V,$n}H!(iݤ98%jJ;{ː.{?;Z~?x_(\KMb .AYf H-pJO+)WW>ŨFlތy8le6ʼL `5J]fN>$a[C"0f"0Dp'%)57v`?If?}v%#ȱxPxl`> .95δ5;SBR6 \:\8k,[s^Ƒb"+;y΄(L 8Nζf~Dwr-H@0zQ-/vݵ-VN:ZEAլ[5 YNl=bC!.Azo7+6 &zK0 yF`7ӆF(i; l`o':[.lAQ턥;6 N=gdR *Cpl41Ų+gæb]E~9lo*W:p8YK;&jPF,ϫyܗaHE}Mt8\ 3u[!!F0mP #ʬ/}n1Y2'q-e KeL|橜1?YQkπ51Ҹ 8]ց0 Ҁ-v* c"EaEV 7|txk k4fN.]@gg@Ц*5QjՎ.'{%lZƉOh5*]E,DVԄGMkaP9Yp&¼b.+aaCTLQ1,WRwt!$q&{@"PHi@˭>8vXbrܙĘQZovpa p%`KI/q˲5 jp|7E2"?@,~sn~|!x|b $9F1Nk@SޣU_ui͎|Դ}Wv."b}.5[l?׮Sf_qD^h3pQR0),%S[tQ1B4X\{ElI);VuVֵNG7UK LNDKM@0EU@AJqC1Ryv"ڮ:Ewqds%wglEn2T="Q.NG0;?]}_^DuD #0NUoV0X51?On/׵v93kq<7$,؄n4y.|Cyf[ PY˟Yqz&%fK&kz9k|t&  >[gft=tgӦ3]ژIk2TB5&|_x1Uq_*Y//)%hcyas?__NWl1_}cқI(hIpRcQh*t&Y@'UhNF"Y͟ݖ@1cbRޥ8q2fW#>Ѻav~dzg~(d.9.u8ه~g?yLMg{7ڌ^8h3Wj:Z8{&W3I_(q3BmH"(cKtN!{ߜ,<#;M@䮞bNJN=;Y>_uz]nB1JW@tzoRMv>ܺ9]5h梻wiDT@6f:ؼoѕGxW@w&c(޽wQH4hNe0wvixގd"`{3 9Q68G_ӄ鷤8iHn3VJXlrףߓd[O,nWE}hu_da/0 }vV(#&CUs+V,ơQFށhj®Fb~3VgzRVp5#9z*rF__mX!CuH3K/=PƪӫufJ1}XmӕK_@dWhطJ+,;$.cldʅ{g fE3∶,#>$nBJF)|g2{5i{3Є0ցs~11_h($elB!MÍZyͥ{WkhD7,L@Vp{j!,Źs˖\ٟORNn>җt=m:g KC #}eh2w^?^‹SxQr /s gwwrWrYc# Vl$ Qj9z^ :Ke<'XӄL򟦹V:|M=mfaO9Nsdo c,7ą_?g1˵Z 2/\O淓?;9?1q +8MciBcU'}Rď~̊8ݾOޕO'YɳPdLd22FeΘ){">8sB :P!:b&RFk4MH鍍Y0`3ZWc;bP;nR#JHڌ5b)0DZ`ZWDJsLy+\:VsgyJsz,0XB]y||ɕ[G`B054dܝ '7HZ!,EKoڰs,9 )cD̖J$t9d$Y H`T=A,Ìl_h\'Ue\%M^1T%K4ȔHVH̐,] t3F+Ier@t8yЪ]-%|ւLgY+` !FlnHl} Ĕ\l; WѷF墯7qӝ i@>`t'|xIK;៻< 7Z#̍3되R|@rA9,d/T&tt&{ vd6_ގr ISnKN$&.)``yEc,! s!tyxfI/_m:CVq!zrV˵vc&9ɍ$դX®\FM0{(Nt6_izR!dI`ZU5Ee4oU+}:+IIVVi[ʩN; jWqv ߥxZu_SrF$קK~n(x8n߸xǟ.o> :2~3ُ'8tqdRu2?!0ڙ)ۘr($əeR@_#>cK>sh `M"d2qmbԆ3E?F* 9yiA ({-D9yYZ JhrL3ǒ:thCJ*Y=P$ L~√[岿Ejد"n_yޅ+htl"5q0,S#maZtB` \TwKzz|RDIB#0i4HM L*1+KIEhi0hvAI @uWwQԂDKLd4ċ@" I1y X qNېoM61rn:N`_K@Y8rO;A^&Wvo7w's5vn&(}ޜ}qwgC6{Yu=ٳw珖܋޺ E z(zu6Wsu._l|.Oջxԯgg2׺fxAj^u7[<[ " wC<j}g[J.%RxHpuʃoKn}Ҳ>wv}꽳r=]7nZl#>;3fƧFER"龘Do$^^ynE*#^Ƃ ယp)3<`?xT<nzO@ ˔=kXdSҾ!V؃]H;/mqucAM':򩧟xu$,S 9Ωs)ҼFdbM2TjW{,%osMb oTc s<5[FRef/cЁIQ nL FzŻkkcF雎4I!@6 x)HLfId'wEv.{k졆<ʆ,`l)#I4\ps$ONsʪk"Z~ͧPK%^!6D.pTHapP\sOZe.jJ@ kٱ|ǃd1:* º |)I"CTZrx^,u EI1RڤYLPGybMJBRf\wܑ/M;=w[@ժJǧ[Q!]k!]#pm4rX89s;s1ߙfM>">ݖQXqS Vv)s8*SKhTS!i>4̙s8)š>"k ,m\>W[ $i+>ԭ\7qUj%u1˜50HzsBvWLUJ֋~g.-dJFWBqTBBZYtym{cQ;܍,|wFM=C=?y!lrfp꼕^C^= W鈴p:9% Q28e#>;oi7qՙ]/:-r[T[r n=PЫz ߥ,}D0l_K}.gIr=`5lHo޽{.{F/MYT.nYhDe4_\kiu 8{8vcBTNӳ/wG5O %#]\#qxF!/2%;fbK^5kbT/ˌ#XթjX=z8ndž ذ׀oXzɾhK|5~su(n<1c2QK0Wz0/{>1s8"n Á0J'ig7)[cI'N 0_0[Be~-!߈/￾ӏ`nZODIHw{9 ЪVwْZIn3X77w[.)A0|kSvk)㹻|3.*raȉcWV['ζ|mey$g?}gٜryǟ_KqjY*e4Zf[ypgݏ?piJS)GˆWI;v(Cf-G\i'F03^m&055uͤ&3@Ҕ&ʌՐRG?&]8p1]j7aFtOڔ 1^#& 8Ƹ14`Lx]hQA-&c:f«+7 zh:UѸp^KQ%fq7b%dT,!f/DKCU, .ryT1it2c6nhի5 L( .0bdr*gc\gЉ{gTjd0Ԥ#j& $K ]t:* h-x >'A̋c IIr[\.av}OM]FAf;=m-%$=itBmrILf]LfyQufVZjv~0)gY+Nڛ*I"Xh<3dD .B1Y +2dx gm-8k-w'HZL'6~._iWya R+$cn *qDPu~bI3$T .Y?l8V[Dڇlr?7HA/;ռlem,xlŃlU9Q_hs@?rO~mՀ]107B z})4Z4i ]nB#EW;ڀ9J̶snnN4Ttx^K(|Ho'vQ%)5Mi752A[Gm)u֐ JZ56(zbּ,fFHHMm]+)9RS%e;'h \I-[OkE#s\Q1B-{HRBh%׽A;}ڸcoGp(p; ??g-g=^÷Ɂ_إQ`/:3B9kQ >](a듙4Q"R:z[Qb*`u8>?> srFԈ:ifJ-);c4i`+tdvړﰯ穏`I5@VnHMpʽvu k8kj;w 8na!Xe?SXuauz2a?@#f>Ig&嬴h`ZާMS*7gw7CBSkgpPd{k ~^i|^s,(O/1 Uy2&:YxsHk!KhZ<} N>z*.m+K/q^[ق7΢mx׺gFoSLHN9HҦ5IJҭ y,ڊjn$:9z]w4?CpuB;M_pN+++1H̱̔V6(Tq(w~sw7܊@P%ʂp ^:JފMaθ[+u!;?Xl4 FNj'е. }Nju=_*?D$kv1ڙ) X.as_ShaLfί_|Ѓ9~w,hXS{ ez,7bZGu$ 4E&^29FT~?۫oH>Q܌,9{!3G8=||cT I4U6r)bJIEBD*n[&uN9kA5E=6u5U&W{:l4#ڌI5R:PGW~ u h;cF2WW [8k6_\3xU_l_Q\#X ַlIe\Cd7MH5+뉾ǂćS 3{5 U"p$1zlu ̓SĻ0 hXjDߦ1Ǔ!W/1v=~K .9i9+h2뽍.8-5- 0u 9o6^5V@hv;VKW+ On\i.V2zzQGvmnA 0$Wj .W=歼I0{m&«E %l.i %,Wc$ºNH{w" % :2z,Iϱ(':Kv ٩x{[aE) y? ŝ1 uQ p|{:YrSoe~8S8/}߳_ONUC4{1clK\O 1s > (bz%h灓Ȩ-&FR1o@MVθA(|_aS+S4soۉ>-#kى1c<œzVk*k-Vض<bO%g,f+߽u."”PXQFDWmT`zôj|M x RҙETDD!zHj<7 x8D翵AOzEK@tJC. C,8#.ঔŅ$x $:_Biݡ"2W A>mA|.I@Ptc |^`JѨ__GwdZהW6l<~˩?7ͪqa9yg[˗D8b,wY# =eJG|_^'-5^Sh*!ApL+SGmTV2LD"vn^jҬ'[A-3:?*n%ڃ%jCӆA LpV~<" K[rYxXv4pqKVQ O7UC|3} gpMp]zs׿|}t6_.wH&g|NfW߹[b[|>KjКz(!BiIh/\o}q۠;! yBjn\I:.]} դ7v89Ocbqv7ߘg}ʵEH{% ݄%'?/|KQ7]8{/a~0|`udzc$6S"GqܞYi(?-4PA0)h"JA% R'NXl6'V~Ȅ UJ7Fsʅq6Ƞ<ZrR6 h{&0vQΠϋ*ʕƈ~0Z'Q2Wh'zڠ9AM Fٮa?% aFrf@lBy!+`(L3DKЯ&SnA% x4hAOx-N$(("}-Cd_ݞ<[Nhٌ\٣HU_v:K1[LqȮ|MrU߅4%o@ 9(ȻKUSe@)dy=UGDleJ:Wx'L\|g_P_3feɜ6Kδ(tQ,^00k#$|!Fˉ! 3KE4GШR#uX|iL.ĠPZx7`9wn;Imk)tP܆{u8hpJ&Or-77-ےӏKIsT+Tyxȑ(㠏&9_=Y+P`S:7엸3n},,Liq:Fz}1̉ hi&?b- srAhbG߇HV>lhFjF۶]=z?֛,f=[v2}Bֽ VZ֞V~عk k[']<4#[عj\cq6Mvnna}cf)o Q=^a58bfekG8eT~⫤rӝi~ 8O_i̿Fd.Ia{ J0Xf=l֊o~3Y'lAE\kE Z-ѲdJAd%rB%NuuWcFVϼ{&kKY#F/rr[~<-="iblE p痆@)K]0͢UkJ Tg5S0a2ڠ|:բ^=bMzVOa5U'gU{S,gUY O F/7Rj ҭ-9v&zV[͕n]pg6<Jui3NZRWcav pȾ稿O3_^Pq¿dk\R*saG/;M?j* 8МRNq8d0ڔR)Qv!f#tϰE{}eyh$Ի_1jtPzXbM S, 苀/PfN坮 +LI ĥB( pʆiTUFTn gVB O8rv>䥎-7_\E:8c?!|fR#M(9NzyOAvgdc2T՘5]8#jkT,ޗ{5|}e/YeP#bT(=wZ gYVM&O.u!Eԡ$UT^9Ijyy)6>Iu\Zfg%dg EOJ8')jA[7`%r_mKԯ !=?BQOP' um}&߯姰8\u-R+1֯ 1"C|)el%Djl˯Av-TJ5௑_#Zŧ5K;sfI{˕c|VbZ1_zqm\x Y[˼] p7:C|fviH{[}KC44`etNP9LR&TL_k%V;,)]F(*s*1X9PڠeG0wAVƽGdq'Ĥ .Rr&ҩbQ6+$%:>. }ª c^F_(U;^T59!#Zcn(JSc P9r ,(\5Eq3& #$ޑX 9`Q4P-ՠ`Z6"2ң$2 x#- oWY<24|l"bD%"B~sR3FX0e@2 qB Ml ZDfWiFϞO   †P;EcQ[S#u4\+ll(0gfo]Yj/6 ꚲq\04@ɑfܔ]?V% p7990!QXkLiFZCʞWJED$}n17J%ܯ0BF0T'YכU{&XZXݲd4-jx>re ̍ 碿g pX~Lr)CH`DRm s}0SD߼mڿ(/:ڿG#{q_mQn[޿AJCڿt4.VdՃa1:bgDz7 I<WOкs&+@G>d9. 3-lKi3xqKsNdCgJ;J3qC1v4ev`ŬlIMt8@[kf=Gj4S C BcK4"ji4`k Y$ ;Kn("FB  - HL i",FTtLް!g/nk>@pX[Ip'0cUKF`5A2#'5E^EQXH2OAaia}gLN!,09)NΒ !K2w_K0k,0cyQMf|mwbI \Y_>{62T.XmnnKȶ)߾yW+BqF\f7}Mg7ŕfaΝbAb42&dQ0﨤!jEL_ g9K"|qWj$s1H!7}F62S"cߩ'va!߸P\M5RnZNbb:hF~Ac[@S[hMi'qy7F%w trĻ`T[@S[hkJ!7Fc"E@Q}~bs^%. ;RS\vDtr%-,x (ls7q}T9¬$u,J6qNb06Z(8T{rF҂h~HWIGE9I*a.X FZcDn`اs12q>4ND_b'+9U~%kmjjZ:fZ/^%AvUP6A+Yqd_yi9w9xZkߡ=ɆosB#n\ޭwq^/V0I1$X^nֿSѷެ[o[*Q답[\d32wOWR"1EZ61;_08xrB\uiUgZ ;M1~5W5_6>wW gaP6Q~Pp%(>#1ͷ:oD`؇1_qP[9r/ꃚ!j)IxU,Q3 ;y얌jkKoLP("5D8oH4[%Zr֌;e2 F]m3HMT!UFgHMF\o"'Ca}nr 5#զ4#fbR&YyRbh!1 M4Ǧ :0r1H1g4n#-2wgOn]X7nm hי#v|q,G6u9Lioߗy[/7ےiڌW,Qhv`uT˓.†* ԕ4JaKJϙچ eC,0`R2?̕CJE9!ƭՓg`ډ\"pzDRSi2#5eʰ,~s+H)x#Ąp sy˜PD(v3bs?|85۞g%mIqW,$s/7fnJg=E]?֡Lya]gpMyN4?A(aVfd<(Bo2d)T0FRH2Hetҵ(Kdc!PIu,KxJ(*i?}P$&F9JxTt37 <+f3 .`hB Lg|7͐V"\M/={A2d_ܕ۴/?meL%*|2yv9N{uf-ip]*ֿ僔$o0Y!&ź63h;k%:5ޗkmT<kYg aZ8 Q[d48$+dK,jX^&b޼GIMp)y}Iڨ}ęvhur*եM] 7+N`{ՍYgjv"y}Z< AK?~v&Mr5lhS+i ry΢+{]kܗyݲǞb^iUSH)6/+X_~ ϋEV^pstu LJ*;{ŜJA+/q&~.`Rt nrF)GJ *fv>QҺ7g5h5\Dݧ4rBr|UnD&*KV)z?Е[[NS^\?vSF)=_j.ӣ$k1]9&G`;I>MidI !#XՓO`1=].|!ޓ`eXV}.\dDWWї=L^РVJ3p,||o,b%,F%):G߮gq6 Q v8Y, LIU)3 /ظ )+WH:Zb2`V $Qb3Ax:`X/v%=2",ZVw3/[A;XqѦ ~%qay˜1o@kcLDem1[0:ĂD $ CMB$"<ݫiԊ=}mUUʆXg=Qs;i=@utk#=es &31bF ^Z Q#|{pbi6o2[2K# & #cv N h+sS"ړ 5-{hA0o,mL!8V?ǟVMd+ҕ85\b'p4A\CI4ԡqZ9vH8wn0*Ȧky,=yqu< `w*bdz_]D Z!*ƈKCh :ZZg%#l5->U,LIPl}n&@P[@АJ{2< S=+X+jm{=Yt&+Tg||&̑AԶ߭#iUoPFrhP]wmo-ET-2tmSwtmoBקޢ<;1SF?03D|'"l L^ѽo QGNҭÎZ% L^tǧ/.oSմNJvƘC-0^FE~ݼ岷 ON_IBz3 9zx.yJ{E&f$]/xzYDJ_%quuZ-FRR##ۡ-F5(82N,N&[qLԍ SZ ȴM٪Cn__x8݁z\*!V\cv}]^Ac𒩩KN967,_5}ZZ ܣ;TQc<ʠԫwYWBSCm85$ "'WAmlQE+ډu]v(vTJZmKl`z9N/-CFV/C]0>ة1`o/E}kKC# Y`"λGQeTIeѭM$EO=r͊-2OzMJ7hOQP&6a!gnl ߷zNg4n;Pvڟ'On9,MM)1aB)|yzrP?]|]m?<ޙʹ?W?.7wVVųt񉒪l`lRO?Ɋi[[Omdb;p&[ 9smeSֽC-n6л t>w;]fԩ7֞'ѻ尐37V6ataR@iZrdlQjU;9-LK99Fe)g M_֑5/f)O #AR9c$R'LyB) )O:$ Ly”'d HƘ'D1 Su$IF')O0 H)O9d4yG”'EP?c=*[ S*uyKl+2<"(Z/c` w7@*7S mxE+.xX<})x?}u>œ[R?ݖbE=Ny[,~)y;\}VP fI,,J^SUkuzMEեaRKAPx=RjBi%A*L6P,[Q'0B ,,¤4QUKB.ǘ4У=0Ά:V'J4PqNLiʀRJo e [[i@=)z#(rfXYɬ!gO2XƹHB:^XhRݶ~s[7OΫGtA*Axm<'^r1h<IRsdqǡ`.2eIۘzECj`eZJ LSk!X -eC4^#\ J00oP:RRv} ij5sWV9hB<~b#w^W-H(Z|g+R(q"/*:laK߾{ [JE$ao؛o>]ܾ?ke_Hu|}yQ拵ke|]Ν/v.;k~xWRRTǿ|_(a⊠T -//>{]MpuS [O A̽׼j#wlR,[F"qUq02YṂmE Ct1ѕXo fTVH+=U=% RW|y&aD0nʝ/𷍻~0w7^}]uϾlTj_{;yBzLžfv|{=x+V蚙"nͰ0ee#)1,UzdleL\ S͢$*,v[NDK_qʥfBqdIb NB[b-XrGm1.f[e`R9ӏr} (hEM':Z81O#u1ѮJg8mx$ihI'[lKLA$\Gc4  ȅ (-$-Rm|BF4_~{2h$cek~:SE4Ĥ3Q*tB~VBy(p"3ԍFM^.Q,,|>q_.7(N e/ N{-}O;EX9JyΝ&.~P>Ϧh)&*fNhT`-ާo'ii3ƭ(ʵOWW IP1s-gf { RRiߛoEPJwm#ʊ|_K/|*ؓVi sYV jʪJ .( RVG/ +\v-axX?bm͢,_]_Sr(^үڛ;`f<.]=D_}}qS"b>\ldI{?ryDxU(PTuNjF-DWVM{nJx| h: Zܵ&9p^H,ⷻgVpV۝l.2P@zjy8c+57 ز,$PHxy'oygLj8vљ.!DeT@NI|Ab@p0\!B#ĠUAyƆvzPZ<-DUT)h36yOS(O3̻oF2vp gӄa:Q3Nǥ*Wv2 c4yz4yg`jέuOաSAVĉuwzZ{N%?.Ǐ0[H{uMf]00g0%b!Qca DD_EOsVPTMu !2.DSm>FXB@;$ec: 9$,B]xFmp?o~|#:djFw nn3rSN.0=ۻD=`kLR3hG@2,E~4(n^>~`ouI- ]M헯E)6 [#Ad1]8dǗעʴRǠhʤ sc4CA@0*U"pnU&,`;H?uqDBp֩$5{ o%BKZXYhfT[` :x0gMQZ9jKA=S /[80ՖsƼb&54BC `=N7ݞE0MFTB!0Rv1&_eb>crrfFG~Np(N31FeF @I}R A}1Ղ@APۨ7"(FV.R *-*V*;TR M?叾7k I A_pFC.K.π)Xκg_?h+AgmЖSJi&lmUtV^ ?FX)U3AZ}w5J$(or1,IW#u)k'A}ILdFMzBC*>55y^ˤN9*E]K( jr@xY80`%N\VW`뉐\YUHR!9Io(m؂dY@-1U~h"ha6E u!0[΍܀Vt[-vX-C5N(r.$CJ4 _(G:UU͝G=\NZpg=A:3ϑWQQW0GK9h)rj9VsjL+SGqlwmXsM!'\$ Tqٕ*`(]x@J t/`9nPA6] NLm^j۹!^NHzd"k$ FT9J$i@Sl+oRe:'(%48&hdR4Lc3bəQ6y$ϙ?~Z_ [tN6-T)R.7FBntߙO}Kǽ4ݸGJ}GbN`8A/b1u"t޹KШؠd33dUfuhԄݧ_}wҍDЮݐz)V{0Fny ff;VTkH́uœBa"HƄOYK&j!/s\n)BH6a|,H(3%Q^FlϬ3#l$7Bq ԸI}켑\fZκՏ?Au٢;GayMl%=c _p-T%Ҟ 8ZӳjZ#愂cUSҔܷ;&S)A\JaǤ7FùOõ)bJiφ{qcV 3LN4:0sP7vT \{ JFyh3AžKZ+ݳS'ӋL`tGڶn_íCOt)V J_ɦtӝ]/_ x0Zʣk /*hgO-53tb:X7i&dLpbO`psI'cÅ'iԥKs(5-sex;= :ǗK6PڡVXZtzƤb ۡΣf)ɲ6*_ƋLtU:oF?niC ' Q.I2jP )U0ADAeIi=[\'8RHIv)˅5펻%!޼EŶj BӉ x};Z vZ]R%JhVC-&ex]Yf º]Vc.s[z>,iLVӕC6~_X] 2Fnݣ'tC. S=~ƤGylrsa8\zo.moO{uWP`:ƭȭ+έ[[O/Tj8k+?@x3 1_(? ||k'AAKi*ϩAUCms`b#_w'9vQiZIJLtn9sܬ}+W_WHسFB$AJGgV1 j ۏ٧[-nvk.ݻ[CC#d!?#Qk ! ź&P%Mg,H"bB85 -`A TI*R =$69,wY*.x{3(8##X`$(ѦsH-d: !C7t+?OqNuP- ;EdflԆM .Ϭ*$s(OrnX%"R2$#2EIjGY">$%*im`5֑YOcu).[e‚`u+{+BfAr#u=`av}l Vic` -,=`))l0:dXS}W1E{bKE2񤙶F5Ќoër2# ǨDQq%`5H \'S`2܃*`ؕrbZ1˻O+3Vܙ=.+ZccU]{Ca6hK Nh΁@MqP%J8 h፣0 EC#Fhm9f8VL,%4[VVaʩP|].qc%e3QPSFWe8SV98XAE.d+F ھ(*04DʳZFFyGa`)0Wf>.a`ml҂} ƻl@.U)G#ah0X ^ȉG|n{@zL'AHaD&&ݩք+%[=p ]W_FADYRҳBp+ #=)02T_0LlbD* !BKހX4Ab3c1g5Bg܃9X-(Z[F)C@[5kݝ`4hJn2 P3;̨i3?5떓JLTL\|Jn u)c+tE6(oh04Kjѓh> c֊:Ip7k?p'quwN}7ASazl'L^mvV\ǟV)U+.߿ao^?|_`6;Tկh׷蘘,+VJO~hU+9, {n& f7Za!*r&88UHV/ۨ%C4^i%6}&|x`b1~(ռ5NSbq)uzXr}8)%A\+ǰdlaعb5{`F{ K4r#G3F;I9x2j._7,ފӌ%T=MQ}?~ݖ_nȤڵ1ӻM:N?,a z3,Г l29n .-}/ެ+è}oo G7Ck/7/<%OnGnd  8r<9Mu.zx a/z(UȁQz_._>+&,TuRARɷ=)I'z}Of_lP*XQt'=MoMyôiypδ0j Q~;0GTYnt,cf_X94>8ƿ|oD7 &5;(CRj$nYrڒpwfqg2v6r_"™i,Ifs0lyE*AifpfYJiB^XkGUw([yE.w%]maQ7H]9+Di?t1<\[?;45x)jE_5_QƷ+ݥSE/Bu?`U?Z(Rui7rmtO 6խ/~<7t"͢.k<-pZXje@!XE$/Q ˣI8MjYMMQ#CD\"(ci<_> Z*Bj%u|e`חZJڊv+5˝D RiIRE8ϳ)cf P̟B8tv(FKBRJ)UZ.N=%ħBlTZ"vL:suCPm&IذU x:W]zD`RVTsPlM?zcK˜ @cXG !/8+> Dxw&w 0rb;?1wM`N޾ڮ xAi+B{!-qy{<]Wy#~O?ܛ F[ҢrEMȜxPs8ԑه.%" ?^s?d<O~m?%Eb08h19iD&sV1!pD M[EPg׼U8nܠv8WQ(>\z^ ͫkMnh4Sŧtvq=.M_ E!oXyX;U·!?GrU%i:k1 Ta!ϐ(i.mdN3:ˬyn4^04M\Gjh=rCj]Zd"_r6t f/\ ^qYqavSL:[UȆBݯf+_AvOnL`48M._ˊmX W#绐!=F܃9lJx Ƹ,@RZW,XKJ4MA530#IQ@ &'J8+G|EVSoGsLCRl nA*wAʴזc,WNQ8e0~lFrKc覜kq!H,d. C%F*1YFxg,Ms,2oe(7z^PʜvJZMS(x|&Tʌ @4ǘfNJgGoѦ8 Ga3Q]-VR<&4-޷۪5亅hVJp*XYe",k緃_J>0ox󻳋_ݴXzdOa'v<~cw<~cXX}f>szrW^޼}W0"ϗa &xj>EgyJ"Dt)pP*60Nh%[v 4N\{S(yG%'OKl(*KKgiV5Li5"~8=8GPME6%XT<IU~4/C>||uM~'ˀ_ .&C0>W+|CxcSg8QI*`4^R`}C#fߚSMRT JЈYB3,0✰J֑,=x޷Âk#f 2SROL\ܦJiF5|#޵1 -lin RvX8ALr0Mc״lJٖ=@şyqºT{[;EKw>2]J졩 rʅDP%h:#FX>l$?R+j'e"0OǵʥຳY}FŬ\H iy4#ƴQA#Dl;Fpddۂ6I'7' ;1v9Qe;FNƉ*OiiZ3:-\ hYNp+hݝ"t_}F9Xcś 8[:0?I0!K{=? PM*fLE L(YitzV &;hO"zA &m\0]LήDDRMI4G!͟O&Χ$C֊XXɐo=B, EZWյ4U),iwj~I, +{ \,2h4Uo3֑u.84.*ST>C.nq hkaoňxNժœtuYtX9dKukx[k,Q,bEUa_u1^џXѫ`HB?/$ 24 n Own*xvPiv oѸV.+ cq=,aiɓjd"PR!}bDn'Mڈ3rYU5 qQ8[{[zιzhA 7ȣϖvЁ'xM 4(IKz{;h Z*7 _*h wc]g88oAzP7V v<ؖZq_F"6zh83-zwbbJ"՘cݽf>87G'dHi^9# .R픴QoʌԊJI錒:D h%׏˹dXˢ" t -ST[oW=ν+G0$^I+s ~M& 0/lou |ko+Äf*w Õ*|dz<7 KNq2Lh= ԍPo_GB~O߿X#䓬ֈQ_Mh yۗ(qNq0Ov| ay}" ej¿kAg9x[9gQ}86Zө˝U!Ae*W`Hs  DPӘ30,&MD ATR!}>\ͼsۊRBg/,pƁt>ކΟVc+*T|8WK;v\[V`t}w:i=-/ScU|?dwgn{<# _?*[ׯϞ}%Ծc=/#1z2M5p^="divluYgZo5ߓdqZv^mge\vHfq8䟜E+xٟMjҭ֩}G֐ݥͤ[P!,: |G;F+NRp;%b讇cVJe~li&y-ieEnn˴e缎muJ&7#^HKj\s Ŗ]틪,ڗfbQ" gߡ#Ϭ7/JK. qx3Dp#q!$󀚱č(&; }7u""pA]K|LYὶEGAD&{#k)Xohھd=^ؽ;d,'%I3p1 1.dTKQ]^"[P\(T $%MADu>j@)UZ%e*(0d |lbW ;å o϶_AjQ*4(? u`POQ9"G5^ƋnTcJ2ǡ@A!&"p4CP`FP @x\:[a ڼ#iQM]t6EY#$K!Ef(! #P-%t EtRԉ*|;fX5GXњE&pDHp˃&oq 5^sPD&…:4&Xj? S ` >UpNI(CEq4Қ}\)86<9Ѩ]PJc2&FƝ}9AĖ:Nl2Sǘ*fҝNOo]RnݚYܙNgq}|-RM-B(DN=q 炆Ѳqv&2o31a&!%[$]rƢª$!3DRx7$g5jm!m Z+_Q)J71 V4 m©EajJ;cc3@"D+egZ88j!;bxA;x)pLLXJG Q!#2TY&+U#ZJ2oqymo}\"^VyH| yYι\/x9@[:%=lO9jڨٹna*_~E`ـʲРW ňFPۨ (s퐢O80xZKs^۷gP4roˏE-!T(A lzhN $xA Y*c=v45 kwwtC}s_V?B)Cޟ1叿V̛1h9WjsMvG){\Zt2,H_oރ1RM 記2Ul&5]zAlja0 iho(:wD 0QSoYT/(8ͪ#ކ)U1SIaYؙM!\ ~ <0b0ECmb*c&t]ݾ3Ohؚ3A{[0jq,<'Sи.LN x2Q8BFȧAx)ino@3" *6E%&]\pÄO!Уa4lTܬ Huf빖RϵmKﴝ 1_uhxnj!W5mj6eV\8Cwf0R>$DuY@.CAq?İef*vl:|tGQlB(a%ZOg >Q2V2VfP6|W34ar@Wv1(+A޿by TfZ7ܨrcμ͍h ϑ v )(xM774zPhٴ|ЉڛeiZ$, uʗO{֣1E9lk90|5FN 9\/ZiO7PcTjR^]7ZOnLij>^\oo{|c/I߸xq^iA:bQ/Ww'\CY}a臯. 1ʆט[!tyDVU ˣƛ # т~䮯by~\n}Q7*\(B; 95GuIEVj4ZJKkZR|õݜKjC i[c$V9v$+;?]9=ɱւO!we >Dv En}on(Ź|N|;vv;PѼ`BdtJW$M=$s@qX;<[gM|2lLI'os(&^Lz-G+ %. }ɗW-TY/w՚4Vj2Fװh}4ݮFP#.&e ki,O.@M#\Zԥ',|lt˜CrJ>ٹl)^sْU%Yr3Dꝼ9"Mߝm!҇l爴!C#'SB˶m=Q[*y߳T4m=m݁x@:Y;BNN I\̋#\$4W)=H h.cnd5f+A?L:~4T=ATSrRp:]]7&+MNNvuZ:YhA9IK0^܂P [ h,yO%[/%Xew-^26(Eؾfi9d"$ZcE0xǬ!rEr3h$qNb16Ig={>g%3B,T\![5}ZlQPӞ]2Ed!rlҘ|ԒI13I؎G$,*S fZ96Ӌ>ZnJlij"O/0ףp4ƺAQAbss\^#//V]U8&'S2&" qRFPOeʀ{Y%k hdnɨ@Lv< m~Cc67MVɂ| TAI-'1&04ٿ(7 Z W8\Ĉm 7{|m&,Xk*ہzUAe:uPUooB;22r!*3hd2Fpܱr/a,}pͬ%d',67('Q,Ӕ"A'郶&)49P&|E (#zs+{! sK1 -7g| ;%0vQYhT3j'@EO ( }QC$Y;#Fn}`sl"W tTaH FD@mㄳT D NH*33z!QV95IF3zes TSb":Y!4Aj<@ϖ@(nY6jYT $'CL#dk@Zqu^1k! Cp <믛ꇝIr.n-{+N[ϑXMe@?Oke!y7fϐ ?_}wMۛ귞.1PVmk!ؼ_ "In,qʒVi}GE[Mi.D3sxx9u%8ihODP±=p @ &M;3+j #;m^@F0@eS\zymEY13 &;30>[L`k꫱;Bsm+Y]HaՅXLK2WeMnum@WZm9(*#|0yD/&8y;@'3 ׺x.=/T.h йݭIv N4-Z".<q|RʐH#-.UOu@tɨ~t\$HF-Ce"Bi2$zrlgIJNVǘrvi`Y.p,^IP(,4 b<D9Қr9(jVy(ǜEoN|$Ƒm@QQ*g ,-*Z]:bR$k4:ҤHnBZc7AҭfcʡŸFhwN2<49 vCoWT^0H3Y m }^1y(Đ_Cb@U[g[Ibzjm{TEBձOmq* ,jP +͵=^! ꣞H 'HM/&l> G7Q, +bT)ЃPO2  C:NӉ Ӑ|`{,V İЏ~?\ºW~QKÀ 7wc+`6֡I$. S<L|1*?nsKg׫,yLKYRXɯ -կ y*ZE] ްnEGnU1HQǺ148eYMnuh WJ:LP8,^S ؃3JRY8i^Ha=lc[A+H-*0߿}%bq-!$ԗɢ[9+_ vwײa'ɝNlW\m,*=ݖ3A%ػx2s}rQQZe ,W&Ϭ$h >o`UZ5فڍMXF-U5eήѣ F3Cg?h !" ̊x'ۇDܟzO91=wˡizey˛r3b-4oXI梱=8¸؉RՌ| pÒJQk|/ [5pvqiV;69EQjL=HCa.ނֻb+ #"N n26c'ABdB nSNHBB]A :ݒYIQBi{$MN!^ͯ3$π**Pc^I'yM&y_xJ q"`Dˣ''EDTMcQ{ID {ۧ+lr8/W CL*Qzܡ3I̓O4SPuOkSP})LV!S$).2kZ7KMC`%y"X sɷCװr}17TWyP;'L K2Ain %%ECnDFTYQ,o{\ gu{ݷ>Qj췥N]T0=|(׭CV=4 GDZ5t1jF0TD-w({H^V+@8gx$fNMB״}?=|vz0Z~c{;|7pĖclݩNWG`K il$tHZf4P[muw(fJѩ9@a۸y!B3K&Sѭ:I<:J>"z 22,CEDQsG4{61m( Rp 1Ƣ\ZZc/p2 J1` |R]e$|!mLlL)CtLBօ\q Dvr^) X` ati L(%M͆LJ=>;Y|1#sM+f-if0^" ۄ~4 {&:8~^ۣY㍿)}Ї<,ϳztzl<IJ(EN} wiA1QS E3tRi@[7k/0JD.-S Ge|<6d>b#6>>Hc06՝ˀ_t1z!8a+*Կ{1 >Y mD>wpoz8\R#&$ :\JiԟrS][Ab 2<7Q G[X6?Dž!XgV `*V+j G1"Q'3eycܖI֌蚬T(cşys> rRXi;$΀-=E\.H*~hh$w6 ̞F0MZgL2R+.hw:q?3$!.!Mu8;n:`` xgRb/UL0IDUG 0PW*_?ӭDR+{kao*N.55༗EVLZ9v5A8Gy]UQ$2 lS~~OWĆiHswxm69?o;^ ]o48%}~x`N'_T}R0oʯټ3I؀ Q"~N.BkSP $M_/s,*KYRHRѐt\HԣNmbN9!JWW}>UMM5uh WJ:g/u4rú );aݪb:Uu;^,p[偦Z:4䅫h%znк1ģGnU1HQǺ,2p7mݪ׫Inuh W:%uJ^L5G5ƟMfpk J˳#gt#gI shb}j*l}$On/9\7?ud'*.HI٘me4eL-EF07h_?iw9).I(*pg~sZ'2 Iep؎g<ZꮥgdcIS C1Չpd.dO;jًDa%|^/Hʣhs Gt3BOnB9=]8PM. u^*kv !+$zg0Rq)dAhǥת"jMxYɧsPՅDsw3tտlA~ze8EMqzõT5 .۪R =]XfV));)Q$ L8tk SϔF {І-*dh%P{ÔE(Z?Ay,m!(APR`@,F(V {Nr Qz'Hd=5&#aRI C( :^΅]CF pœs0Y SbJ4ǘ2ÜXċ^Z)cx;_>Χ&Q]%oCj_]s>%-X)e !o^]0+oj<`nn&W}֣e_{`:x>g=)ghfoWRghL tH1($CQ.nӏ7qM&*?\{B՞qX\(wjG{`Tah4W-7JD щڣ/}3ݰW@Vf%}0W_65\'d%W`c?7vjh} vb)u@V)?^bb}*tl'*0Ra%&ZHLHh5 $C<< DwnMz^1Xҿ}Hx7lC鵽8}Y, 4A Բkg: v4FPJq$BpI^ݎrt^x`6Ij猛N2nsEΙO#*yc 4晁fW QdZ."6ɴ@Ghd nEәc3 Y}C'3 PU&LIT"FFDQP,Htg۟#M}/n6_g^RBwzCp\Xڽ:DBj33"tm3no:k#{:,n[0yC>9zӑT<}H3 N01C6O>Bmgՠã;Vx_+Vo'=~bv6Md.>\?cuTbﮊG }B9!yYYC>.BI =,,ܡo*-}\g p3;0Nf}em]'(WwHn7[,׽+mhCl~k&P>\NQ`( jw 1E7s3L틨Gi¨zUJ#'1@J)rͳ,2+19s.3RKiGI+ƻ,AH̸T㮡OQKԪM;uDԙ*1-2$VKjƝa f\*J-[Akj;)2 aZVEbODv9~"3iֆL)(o"scqxdb&QL}}Лy15cWz]e{! w6l7çgwvϷw0EC7?u;fh.bg wr=դާNK0?ӛzx#ƙtw]]vZǗ%Ƈ4̙Q|Xm ?kpSx|}{pKn3΍z(7-8DЦWdĺoQJLCU\\wέ_=6iu'ך;$?5Ŀ()q  P|÷?-ہp J݆;]= _}/NA4zG!Na:]C6&{?r?$h@ S}g2ύ/hxvh5a =o%TNhmU//M|:vƟZ`B5Ai-^,&O4pnw]!~&M;t9q/x~\17 _dھN姛fk%(o>4h>Y1V8X4)# H5,p ł#wC1. ó˝sv!!'.R񹨭|"pwmbkQ&euT"O,vTa2n"DHYys>} 3EJA/k:pdu5h^4 -\S,T- kFP=] ť1nFP8=y2]u{vZqbE'otv%'oXttζ6غ+Eq64Qׄvehlh2^64X Jr}}sigQEF|1D#z16 @Xl*=;pTHd2z8]@<2gmF@M]W~_"\:3Bŷ/F`E={ͅx K1\Wy$+狲2X4Ob,QD}J'W`4]hH)꒤HAJ \Δ&r) JsR|\b{Lyب}ۋ!q3 ߞ,Ԉ"Njb4*r%]38/^K-=J {ޖ_y[>OhxDaܤD%,Ot:~[EG2 ם1T6?^ 7c̐ A7kSS4iƐ͆;^\gí>4Xzrap+2"͆kQblQcYym8Ĉ$0&9A?NŞ ҆[IJŭ]'ڰPs$8Hvhs! z/tO>Ճ -PI=peRI2deʐ2aF73]H*˔Wy@&?56@pۀ; \\/meη [i,Nz.Ci*V%6uز [W͚1 z'!&ZuW{\KCA;΄k=(qsw^e{qwF6͍[LASeLSJ')U*aA*/8yF'tpq1 hnO'[tث//=$/4czaKKbl'; 4=U^NPvp\3 D+Njt,+;OI3bnΊ}8iDP3`L_6 4#{OV>Yyd+oneVUPJTRRδFB2$O e9tH+ }QP]ˏ`@Ю38|tOG(j;79c+)UP/j(YItU1@+bVhp dJ,("'xyn6*~SըgE'w2&Xz2x}E,UĘ%;U/YbY`a8WE(g PSٮm[b#l~eTBwF Q,|Sӳ-т/) 9 mB{0/VV(^21꼰2} J&\"MR)Y"TFu+ujlSq!lgaGӉB-U[rwmGuXD<ُ 'gY)ˉRaG8.C1X?rG7cHS+:os@:xR?`ŌX#5?K#j*{> U~?qH;&Mac^Mn'}lF8`v"E. +|F<-\ ~q>7[@u'uÃW:jp2 AZŖ" D*+n7dKe& ԩL vF]#xdlwϋLJ6;VMz%9T '\i"7fUZ9ҢßUZci('1q6q\%~{y$l俟jRP\8,ˣNq?P ^]J5.ǹ`TWQGB?/9i;<GଓP!jwsUxu&/)|΃*֠HnE:N'%LWB)Uߣ, QX~s鮑o_-U4Q:ϱy,Sݟw.R|"5߹(7߹O'7cWX0p!MPaqkP#YtY0K 4oqQ\̆bS,Mdi4 '[I LɧSn饬r ^.&$-D-FR^#]x ٸm=n<1S#J]+^QH9Z4kz_*J#% rU+hqѫ?[|l#/O&wLTpp5qπz8s  (fn˫^[$n2Vő,0MNp2if M. ઑmKb)hHʄ^2}lv co0&u|z4{p^TmtdZ/./FǷsV+&דjA{:q%t~ψşw.lݝ۞;~U!'(h(_}{tuw,hA6EApHRoO?\ cra2Dͩ@/ޟ/҉"/҉|"/vu\x#7T`D2JJƽCZv LѦ5 m:yY>Lڌœý"*=Әt;1Y Ѭ2((SżBjj$`"zDۑiJ7''Qp$[Rs~w`{=9 <R^ly#vtbme)jS&IG$""JQARX$MѶ j&*%{TU9`|A!ZCVz%8`K>5mJpJ\Ѫ%8l><TIӇa 4,{-B8V *PEt[6EfEbZ(BL`0q,L^ H? Y FP0Z.!5 9 : j#IuQ -Z1'7Wq^/ ,P)נZܟ>@o[ʽGa5k]y^8#Ϥ M#fP הY8p!kNp0Vi,TA F^ ڢXV/vlh-$_w⎞\]r,b68i&u~IY8&MbyǍH7g^5zs?T]iZRPDUViť)ʵPTf A: '%RDV9)[XJ|'“`d2p#)HymEQ 4L`DŜ5j1vx3*iSY4xb&Xu9$1I"ZYR?fjpVwIqCH voǀLAjN P$;PYRL4tclÝ3@3Q4[r!NMJiH+LSE:r(QI""/gV .bD=93Mrvgv22$;E LS{U٣UQSW]-bz"qdl+p"voՠ mY"&P|w'+lCdq)"v|Oo?{#KdPv&MJۋR AS&OK=v7gx*߬WG#{i57I1 50FB WW^*aċxTxI@${/Ut `WuЈ1sob˗7xFl`_R0GdEK[VUL(KQ䲤βZ6֖ʕYIsk) ezRQ#ګ]ؖ뭘ѝt -Xk&lrWZ* 5##] 0Bz$QLU Wn1\ev ]s`` wi[E=.Wxgr1)Lj=r HY.GozKٝ?X egZ+ { q`x! (0F祝ZPgO4&tt*vpj.pڕ8DŽJ;I]຋w"x)Y; `qzbLZvӬ ߣ-S÷Kin5a=zk+ u@GV%Y$R5%l3 bo>r(Iyۅ[k! &,͂w9.f!cACng 'xx 4ﯦq xCAռZwխ~"Ƶأ ˪1ؕB)_Ugvӽn5moQ랊uǺ֋7nPٳOuSSfPI%UlpiVjaZc|P?Kp@ǣ ArM'+Zrڤm`+͋z[ Nɡ h^?*x#` ^ZgVijBs_&:$:rplx., &HRX/(X ͂~{S-e\]Zq7-R&9x vj˛ [28(UhBTX<$r:=?~釛ir8p>Eogon/òr m'guyxT)}sşM-~l>x (i,_-oL\_ӟvQ^:fcr)=KQJcX9[_;hRCc N>vkA4v;o<-uyZ5?fj6$+F2⹝t=oa'pٝW.Gp -Tflfrih*d]h]ܯ&aV͠r0:15.śЄj%a="CXOZ&ףTvRpGj"kM|Nzr<⛐ʹ?{/6)XFzQ%Weɳ<~9 ݃\g;t~ꋝWwSJn7~[|-,ײquM!ll89$hN FVƦ(g7Nr6N-72 S<~(A,V<<8t.h=R~W$e r|[o]ՈkaŽ9;TJP='\ JHi>^^Ҥm70opiSr{*1Mds6#`0[XQ :^)XT`h5X-N 2SRJy+ 9狱gtX8WQ|33I4$gj۸_a^j$Jvϩ8)RBQZrι c\,1@7)-v ["p02"JFږkԞATU-n7O7,( C-኉soN( .T i hFzLg&iF0`ʑZ0b9H"y V`Dc+T!_U` Vg|(]'D3q!:-^ $tGP.KARڀMPBZ:r⬞s+?cS/" 8ſko|zaa+>KWDȻoWk̪&>onADF?v lث1|7w&f6nn~n )<&[ĞJ>rM7P #5ofߧ7qkJ`.h{Rԙ ꣋3ltAkß@YW㥡Wu0&#a%('WPoeAgLq#nۀo6[ ܡa!Ur.$q˹7aM*U횯ΉUGwCSLuw /H_^*eIYn);t)w3DYx"[Tq4\7)0ָ(em2]}MB^&=owVHvhH(y<Ǜy2O]xiT 51w$Oyk4(K K[-nxU ;L;mdѐnY Y3bN1Dx띢X\݁Rjo.nҗ,/7oi]H *,%s"禑\>wITfY,q{?oCM3`\kn!Jט-xPɸS[$YyE2Hɒ:J⼴J.m /61 =/}յ+FQL?9M7#MGBH)iLG) Ǎ%^yJa5<}-9}\3bE%A`pPgxLΜL0?j30mBu5f/bٽFz[ γR͞-B̸3y>S ;4³A<~sNmm9jX{!@MY5wUT=§AuD/GOJ"yd X}a81!S0RjJ:r7.hY' #M*E Z(ZNKޠ)E%;qH6%]. Tn5T:v+DuzX{E?|P}U'b+lB F+teaRRj+A;Hxa8/`Q{?/,%F /"o}v=yPl_Nnͻ9߮#󛹩良Ri.!bZ9#'Wj&R1k0z'_y5FЬgYbB0rwTI(췛r^SXq-&[srPl}SrB _,7[ M-_,)xǐ!c 캰a=z_>32di/ Oy:pcE Ⴊp4ҏҏӫܑUzé 1*Hר+}fU[i½U kh:!5zZvN/>l߄@„@y "*W'PT)iմz)DC:x4jUmvUmvUMjPĤp?IV-L{e0SE؀Kpz7Dl5UãIx!?$=FM{; 9';j7FvKu):V@S8qr,t]Ryړyxz/ÑS7byD+P#+?ʕ!"4t$VjY K[-0knP鸥2I |[' E;^TlMw::1{m@r/Z"7Ssì*IP!1Z]mj4`fƅ(:nlf]W*9P\a:!(V:N%r/ܬU &|f cNQUI8^.9hb6 ,TɤՌaE &Anʊ|ylig(Gj*tg|ך-AkLC g`K*ux<.d%R5=Ì-[~VhF1ʽ:] Z4 I[ef JNbRY'$HzE*SCluw90)ܒ/yHuކy>'rۀ$2R&NsM"!|ŘOfA>$i%ƜRm[GL;Gi^?]G[?uyP&OicONj[<ݻRR[iOI8M]&)S픩+V#R]')STȑRKӚe4tcbNQ(tvu>HbwPU&綦ے~c>ۚB ~3(NRF}YFN0Rʵ_W^q~`|z~>F󇿼_}\`4ewݷ"JɴD,P%WQ R3F?v lF|w߱xӗGW`ex41wKgcOsB*ċ"OxtDڊ^5)U$Rn'/; ǣb FWnkI8=`կ\خgEyZn M6Bol5I7bY9KL46'4 D .f# ~SݦԺN3T.^Z.WQƭKhx . ]< $ʇlj_TJo ?1X]YYsX>Bwk]Uj6FWԩL_o׆9eP>ӭpo+'9{a= Sk¼ʅْ|&eSǖ -nn!1ebd:h4,:-P*uvkBrͲZh*Yο )Bk뚐?$e.%).J˛H)!-rp'Ď,L?mb㭙OM$gFwDz8b z~ kGoLf64 Iߢڂ*, -i4=y [UPڵFݺQ#S:N}Ƚ|6:"H&j':E-|LԸR+qh%>ww_ڠmZU5>P礎 g6ecmx()mdk,9P[TGoc :?ƔEHoŇ*\vXU_cOV\S3.%Yg等\zq‚="} g|+M%.=M+/v#!HCYX &Pi|Ie^llX^"Ң;c\ y2I>B?}4ZNVFz}]J/NRŽw:jVa7Ԡ`z4S]S(@T1}_;c!;Cchy}]>X#d"u:\W_IU3ҕj kkXC?~{>_'2!Щm|}u@-dmOK^ v1IҥJm?.=BEL|{; &&N>ޏG~ս__/?ɍ\2IW_.ߝy ru> *HK\{ 髯dkL4nDLL p EjՑMV|.qW;!D2yw{ެh !Npqxh{ҝU| ["_5ǜ4\?9|@ ePP(z G#|k2B0 }fexR yäAeASfNaюIn1qRJ҄i%-!Wˑbrf{|B Q_W:10J# "k 9`YYgJ K:]Bo,t9^ql8dee LkdӖxH%,{he2 B;)ĉm )׻M`2m]: U5K=.>LmwcyZoiv B)4l T\e\ #F *,uPgY:5]vRߝӎekΈ8$UP8gC.BLT4"D(:D3:?d)Ss^3 M^ 'ϟKCɠ?)ؘZ)j79_꘎:1t09]\.4mѧX$or:~|&G.$s|n1x( 8`GRZ /Rsur_1#}zV87w@+ٕGv1Q#Ut3T+GhK Vucn:|*E8aО8νfW1xVE*E, ʤ1K8J Оt@6*0VRz0[ɴUσ'Z1 ב,Zb(n RCVZ}> &e]Js[ b&`#h8?йs-AQRQ7WP*C*W9ɩܦ3P֛ȽR]ʹ0 pu }MWWBwT-Unu)qܕ0_ݰLbպͪsQR rZ,>c̅S u/2a!߹T#WhŖwD)xR rL] 'ԩϻhwBs=ަxsxq@ҷߒHgix]FqLW-W+ZX>0 ~+q'e'UF ¸ʫq.Np_KAHx(x{4SR,Z>#Yj03S3% 4:.kdgwQ}̀3B`WC JɡmH.71X! [ 2S rN:z{vߝgWwtAH|,^xhܜ]_T5B5LHrrBDPJx i Jn*_P}A'P  * H{rfћ,oN޽ڳN_La85ʎ_PJ3`ٜSw/R(.gm=`bKNZ#) u䜰z.q4z ,!Z%>UqZD#bXLT""Vy r(QJTOCkʵ ^Q[He<FX TCѐ>'xBeFP%wW<_ЁBLB|ZF8- > 4LkҍB5plij-)g/4=A=gr "z>jօk1m (8Fcˏ_:FR@iX¬Zu_{7Y31nn߾xax\e=>"qy?װU~۷L~yTMa{H([,Jm?x}<Ei#pa?۫: uoEFO9=_0 㦖-qvVkIa3M6;dz Ch&dfF̠ u|\{ lG 1[bqcyB1 koU:a-ĺӤ1>,3c)lj>xgLIV\3U@ۣ{>d5RisRY3i8K)!F{wKFiG}$[2S$J5̬l&Tr[UܠqwyZ7v\13 b +~^f\l278 3W?Q?<)lƵZ׺hs8'#^)r .*Ĉi(Ҽ1/i^miB'9fFCfsNGmU 匲A.\]ҰeǓe֓01'FFx b?`iJ @(7sa2Rlj+H_1cmN:ct:Sܭ[e qZLS3ceA}!(c\xXjyDՋ{r,a#}X]_+QB=[B/_^/fI/Xzf38pTKamDT7$T[$C`jNpeع甙\ s P^jk1phOqh:RWs't#mt)-1l999|@6aZ*bK?RC QsJO8Ts\PLpS6]3&?J..2 f!^ُ=0F̌V¯lQZ|n[ 4 ."Lt|Fi^۪svg3ZX39jWq9[-exZ'c_o7=UU['7Iw=eh~yvy0HNN]#9KC;qފJ[j~u+7utkurPb.k4Ԡ +6Hx]mBALtW NF|I$ amgM }e[D=kC&3!F̖pL'eD'&SL3'pN{̱,b$G匆 ,O+#q10k+`$aBp uA]IJ IrګL,0:J%<C 9A+߮b%R8e/bڙγqJgwqV/@6{?o4mtdTH5J&L Mg~f*fRN0ETnZ@M]p W;kw׺ *4KJ<צ4r@*W]tCq5TNi(̆7Hom9,I8p6.éÙ5C#s8\ #9(-s'[n*e50,쇕s@(E^DOkUUǗ>\ctgF] R(/:gR,TxS:.ǚ?ETX~7yH [[Z+Srm'JSQ[He<FX ǣzᏕomLP x mqk ڻW3oX^YT{ Ԩz[[9+^ܩRy#eV2Q(%1RKi8UUUX FRJ% zԾd|`^\bVT o' IpE+TJ!!&z/"a Ë-jAUxAx&{{WgZ0MK'!x 8:PzgRƹW/* b-UPZxr61} #@K(2BeE+ Dwդ7d9.A ZvgcXIWqү8PZf~J2BsD;딂(-Ab-@1kg#dO@ Α!({#)\3 :Wt E@U<ʠb|*/bP9hj+ D v Hrܠ wVknHiFf~)~i'YxxƄ!D+aUN[ d0l.ꅛ嗙zuu  ]lgOO.ڻ~꾞ϯxw 7a!Wգ+]}}F_**ZeN]\:"[dاPsE]imf:yrS. dV7 3fq6p17ԌT>j}{۞Z4 <Zf=%,VT4 ռ:tZr2{ 9]?|Qkr0i[)BaE0Z(SK]RO!upͯe. /~S,Ť r,&'] `XvDg8YIc~5sڇ, @mgJB:O#A z򔊝A gtT Z2s^ERE%ݠ2[1\ǭLsk`L\SAUԊiB mŠ (Ap0Uϭܟ{HH{}9 .z^ ǁ\s@$tB1厜$;PJ'?޿u2.4e6+UT2ýK_bEV.n/ig{wxqLӱoοcj^lO>c%yz]X?Y}cDz7jJ߯ !Nm qg~e=+SJ2,\ed!߹T#h׶wS J11gtn'HP܌ӥ3[6%6&9.|鹄r{k/WGg:}Zdժj 8 MYs>Ʃ],.ooJSȿ h\Ѡ;2J9%{Ekuosj+-)O10OčTu Wh;-\FaR:(vYo[u^.4KĹj^TwUeu/ ZK<O ,E+֫U^R%mO,lgg%5R%sT;Obh^ "ڬ%\/!$.WrYHsRN'ڥt+|UG\iT޽?; sSU،geGJPgViK|Ӛ7 "hglM%rИ2YkOb!Cyw W (j&.·r@#>[r6K.m Gg'DCe묧U07-Vv˴&+eR*f趒wJ [e~,ꥸnXl~utJF=};n%_1g2udoβu ՂH&'3bE0q>WD:cx!BB0X_!XszH3g\zJ%HcbSy% B95J 3p!;DIFhJM+)vo"S%ǸFD$fm<^x޸n%׬=oɚ`bFn׭UU w;KJ j֙Xzu;gKyz$+48+J~z-l2]4cNLikB"ɮ*5)eXِDTff҂#]X7[L9pO/ɰR`ef҆c%Cefѝ:Z JqVfbB$ZH׬3>f1݄wE@sjBTu2 7l1mۆ0;ӸTKH[+:YXÉb,B:ouJYcJOe,|f8%Ӝjn\J`(!;瞄ąq 2SMWD^AV-34ɘw(> . s0L ':$3K1.) oW~hz}|$3ǹ6 2/N)Ο11oc2.^-4V[Rz*[Dk9D_p54~ER",Y.w2B !Fц"eF "Ε,"(ISƇ2tFP2 p{- R, p.e$+8#0Bxp%/ِ$ vokP_lX"zT5ύOSS u;0eg_AMFHBUOx13bE,xE"LtY>EeK8gd+ 7@OMq>>Oon,>}=fdc]> lB5'eiBJUfN8Y\,# ;Nܕkٛ"#LuWy]HO#uTp^b0نZ#ۯܥVޤvP}$ֈŐ{Ufk1m{9&!Z]JNk}VAr zB[Iy#bo'8z|>Z5Ų 5ӛWdZ *`Nj5vٷj f~*YUף8Y--&30)YpOE$%-A#3搓Ҳ\x{*tp5qyeZcZSzE)La1 [SAPBcX"< i-6&C"}n 'yXi}EXQV*% &08}$q(<i`Z(ÍR(uP7@z)7+ 0z!!گ3e0Yx~yaZxtGڿ! E֟#>}~Sj0/v拏~z0wX$+#_ b2!Bߙt<2?x63jDn={G~ 7K-N8v$0u\B'G316,|lH}tp})n:Y.ΒZ=adɧi#"e]'9NSЄ:ɷR"xy: !"OL bD\,!j_6HtG8 2>XqY@aNa3QJ˔.;ڙP/7N9ݨĠ1~R bwQ9ĥj^$9Rp`^= 1R/'SZٞE H! .ӹ8ܜşR=ÓX!P>_I]?zQ(rHUHeW5JN(Rp΅KG*XWAsZb#Ws=ZԔ(WerƶFLA-sHeʠ;L2lsŴLR"8ۄD0!7K$DrA.Jh7fu7 XbKIBl_y;.+81 rENXWTZ=@Ar4/95H]ህk@hg֝o&TsĮX؁WUC}ח9DtneV3"6':9D=Tcsũtg*u5*fŕ6|)f7ބ }bYmE' k^b9Tኯ$>>jw怉v,6R{yVs6pZX9 |ՒTzI>0t)p2g!=rKL踈7`'H ?_FvTz~=kYu_7q~=9 ׾q" W{EwyW%u Zd(toӻSxο"7+Ot(t9{qeGx2ɧP̧5fn,hgpšeӼew;2^@w | V򮯍;y?/ 1 9Ų3Ӳh\ϻ *mg;dXdqVOvw>o,:,V'Zx5vt*2nG'Micw3y8* .ժmr04GwOS#J+(cGHAqS%=e΄hv~fiqFU>|z[K)?4cIyعb06RHu)c)] fa$؞OKFskru6~Qbҥ yKɚ" /cCnM1HY:MǨ#.hRv-O4UB=M *Z?S?}޻41{=vul(Ӳy%VΥոafڶV2kYLӒ!Jmq2r R1-74m#N/؈ vn:ʼ2@SP\a7E[oT (?ӳ+S8NjуӮS8Nօ; Ek]j NǎnߒHe1PD *%"#ʲûRԷD AU=\Uq=2Jsɘ4&wԘjuP.wL"Lf`:Z\hQ Ӵc SzHq9҂nБ-3s4F`YfsdY#'qǥwvBc~kƉmj.e`X+ \Iw"JqVE`Vj`^Ī@doU5.bƶ0]?cM&1zc[+ B]dmB%A m%&Tꊠj;ЬT&+)iGem"R a[4;z Z܎˅hfx^?)>Sys&e2珓cԹnl&}atKrz4~u?z979%N©3DRwț~&re\å2 iz ڹ@!B.PA;Kt a[|OWU9=!c"$9 hٺw&0:U1ݐ\G6еm/>-ԉ_37JnY'sQJ梴:R,twBɾIҐ·3Ej >4tw}$CĔ5 3rCĖwWk8D2Chw/e/I0W1h575Q`T{Syᮍ*bfM55톿&5n>fD]8Z[pZWrkӢwnmnu3+Z0stfGݺRmVu8T9Wےֽ30m99ʙH[Q2=N /MHӤ=r/p *g9qCYN=2+Dnb,x94l!a4 "Vs!(b6DN y*9DIP':ӔKvgZ8Baor|@`;v/Fo$0( $即ۯ{HMi({#9UuuWu=~a q܈p3DP$0{¤xf%QjGx%bYL`EnSV`7&UaiLaZG!=2ZX!XYw z2Ty^t%9b.%Ԯ``u~0|gMn $&#e|YZ+YPޫA(dHـl]175/-ta 4;f`D9 MFێ[|ŵQY wOl rG6aßfFD0u=-s~0 4eW%w8uWTo$ w i;pK CH.!D/8A ĄB9fE-ɕ"&?$o#frw.J@y{,\3gf8Oeo1@9a#qFӻvr|[' -3Ce.{ۃF0*{{(>bed]@ԼZ{ 4}mBzۢØ O]f',\vɨЁX0>TV{W d8Uwt! qqrHB `}6[Yr,lah2J9 [xw  mK J\7'&a~b:Op(;({D=j ,0:^pJsC֔9 RF4b1{#8`{E'(-N䴫SS)T3"&1]9nCp  %iɍc3)irñW#{O AyY#9Bg*^cdA؄) #ʑU,]@[1TB !2 ᫞ k$T"*J`E#x%1"X&Q/O&(OZ֨TXTmϻUz=kNxg^qBm`۷}.Mac? yC ,,S3`HswnO6"T͛?hMG}0U`FJjλqgjaWY Ҭe̡Sg@[X#9S3ivK ؅ Lک+Xy "J't/gMI?W1 v#}6\.*4V 1#v#uz%E`lH)iLr5kv4#1H?CR)m)ٷ Vd^wFĥzV 0(s%Eo 0X)Ӛ̙U(fɛ4>S7Rh&~vy>1TlՆ';SJ5f X- /ް-9'PeR8A& 3Xeh$FX,4AR&q%ØK_%xۉQS:8CPKeRWGa^wv7Įf5~鬤0E|K,7R"CȪ3Yx{NtTKB](?-c ,9s"HrvK:jՅ=һR_^fNգ_Ed,Z=`-|ܡEb7gt5OY$w>?j6[_->MAS3/Z$1{ V3 K$"?c]|XtIPOd- fz)S(5.b{Xz@e-^[\Ϣۥa}Դ82v:iLVCB^TR_8b>3 lp{xXkBcc'(P{QssxSww2e6u)M5~jjljXǼ$ܵaf{ pOE"=,~uM=I?>zĺ+Ӧ;PK@S";OR!FRHF[6cD % >sㄘ{gVzYuB|:0"Klyƨ5J0c yw^21{->^9D"ʨq,});zpMsB(a= D a4BَHA݀X xͮXO@@$̱H`<2߹Km<)9a⽹!:@=lF{# `}ߋńupn4ցEH{]oDAW?eʎ1i>5EIlHK-rһ@m ScRI<9DJ(C13*qXx+0Mz~Cϓoo3œg/0_ 4S) z<wTb'sStN~O9)uN`fSwxV=kG zr-FPj)ɻ$&m4>dJ&IVP_S[DU&gŏ vڼD`G!&Njތl|5RO5pJ<A3T;`!97HּZ4 t}v&-1b񵱽Wf\2 f m6>]o{1uPrfɰvB+t<]:qjwskЀ8&wۛϏ -(G?p @l 4= 3??9 zw6_gg_igI'Ι/]\"\ l-N kDoO_BJ $^Qw7űJZJv 6 lFQʃR0y6co2c֐.2EoyE{2M%zh *wrDLw'[U:m7O hi77oMkΎZs~=Wb+4 2?;;y㇠f|~aJZfS3 M uU, k݅X+r<`%] d9w8,,O;'4"T8m)?U)k-C #?$P#Cz>. ؖ0f!F~fa3C8%>s@|[,3w4^\]7Xhr~/8W:`0W'Ħs ˫E yx:r%>?O=BǐФ~c"\q pDDvQX M:̇Gܨ$*?)40>r~h2F f}{,lH\[):ѡq@,q\;8c!dkr 1d11 V {ڡ0 ](QANgt)=Z)rGfAEH[->ʵ6v[-ݣ'{=Ru3ݽJP4W֐.њ%y/&5+Ll*@[b g$bS ot"^\$i㑌j< !rjQQbкgu<"{EuK: x|K~t_߮|0)=ޚ0LO5z*S6[V-_ g=si5&b`% bģ5&p\( Xbţ#@t%2ö+]<D03؜Jky | MNP CTdKG4u[N`e:)I}/|^zÎ1)p,%#Hsd%Ur 3*p 23?Pk5:Aϓool?{㸍e_\|? OIb n̢@T'GlWd%,e9q HT9%y$ ~`Rvwϴ}r\jM>oփZk=PZ;9P4*W!rkrA ݍ45J\Y3XiGߞJo~E`wf/h_ml>}3{Pv)į%Q.1.ULq5DJ B4PC^h΍iMU} nUZT`v5R!&lBL#QoN?5~s2þObE6V68;ys2C"zI6y[CFؖE42JvK["qk̳?rO.*98`id[Osט@ `/!yoʦ<~D§@*`\T$8GpQT"iS˹ !;KZ]ncp}z c޿p:&iDэ! LNy{ITBcW'|ځVn!Gh+U7=y^@:ϑ47O] /WI"e.Le~)Cƀy$0Maɭ.BUK$uF BN-JZn֟6Y- ^ߝ.F k[LPI bFҽFrau>h Z+%ڽaX) Č(R1$JsƋ\rH(} Z æQ`n{ q  Zӡ/$`:~՛hi՛hk.̫>@>"94Tvo 0r`T #O{UHpcSW ;N|PiƄ~VC?rPO#)۹py[_)W`uS|#|n᷎l>G趺ug.@+q2E.$5P}S(e#PM!\&:#=n$xku፠SOFk!SV . ]S)E9BFarq2n6f5M@![ :a+(c`K|D1J4au;0(MOwE5m1Ʒιx)3YǨSM}0myߘrAahUW ށu}dVP/Y*,~>Oz =͓%nwf6)7I i2¡/=DA۱x!H1$|8F9<YLm\ 6zcA( oJelBcݙU{;]G*8?kWޭ͈8?j弴ʈFGjv  _uq:rH,.O{k,cjˢOfOZaˣXtb0d l ):k&1>}jK}{=e`b@CTppe7b/E(_~ؚl,TBXf #f9EfdAɱ7uO._竵˫ڕVS{>ruWn(X|֯spSwmB/\'ꋚ ^?Š-cF{;fy |*(B g{~ O$; ct6V+J7-JCtu'k!&VZ+دeVޱM&ex[̸{t"jEٿeKWMZ(mѕ~ \Ҟ-*Vϼ;Cޗ)tL5|ޣOVerxE䂝7n.W2=Yl~y2cg[o f?S)nSɜی\S/~39|靚.8缞'ϳ h"?ntڪ(TkQ9{O3u}( ( %9"h5XrMA-$s~=4Twwn2/wS-l>S- #t gZv殧~_rD5pt3K̮AlX7gs(e1ߘm8QC6Np,KWTAHP },onL4G1İ $-qRɬnfvlVK(s ޝ_9.Z&Y7QZ\JFAb6$׆2ƂCY,}JshK鳃NʖE@>@Sgu+wBt*8O%P;  3EIE[BZ6j ],60BfOs Y1VI)[s$=^VZ!ZaHj 8p )$4IYWeg7BlsVyY:_?3DL Lh\t2{XN!dyDwABJ#A~(YID&ޯ!Qa^ gR7c!ΤIc] c(a(XA flj%;_!)>z{+v(keXF$ ZerҾZyW 41aU-I8Ŧj`2ǚȜ12 (e^L}!S pd|Ϟ۩$q%9T[\?@O /^Bgjs('"L>Oű?ec!L/u7H>A5ۤb@R>77M&j}|zt3Jϋ_`JjxRzD8'ߏMRw̚l#,ucԟO_LB L7pF2ޜfJ`)jƍ=u ٭GՊX%Jp:P-EAlozTpT<^|==Sr|F^eJjj Q_~=s|OކV},uO~$7d,-f#s& P: h9 iR![h܀p]x` -mAFdo-= DwveDŢnt%gooyLP,~nLPNnk+'͘k񆠲w163h@#S2ulP $W DM]ѪaVs(Ɋ:óS/ja| ٦!Va [c@?dN&gr2F@ns?b` ?f0Kw3ރ;J /ueKff02fe⍲@2E>N(l?#`0=G73ȼ*OjLqyQLiYΖ^ٮ!ϟoho>lfe)1(`S mۏh0 z .HH]$y`)0]H 0VR7/E; /Gւ P(Eb89qd#} N{BB&wY^0+{ GTi 0M%8abA)QC;y#y48sj)~wL: A,~]9 f ς3:^h䯥>P] J4 Yek?\Y llvIh=4~V7PkA}}T%l ٯ%WnT5U> ©M+ߢ"R{ݬBbvQ@HF`"-:ANz]4GX8jV׍da~ dҐ* T>$,hFI(3RIɅr ) PDE *0 B÷a5te؇#$*IP=N|%Ge7$ IXK[&J]bo SEqTғ}?^[XblPa*ce*9 Yd~^(ѭ!cLjg@BW2Ʃ9C9) v'u\_ei<ܕDP"3Pr)f2"+;Y:pl5d?,;%p^θ\x^2V.nOY;y?_<1vg:7dJbx#~}S k}ͳ{yfh134y.I^ODm{u(>+=䣴\K_C]4OUƚG9n,[,!]']rcS^9%ڻE74ԻC]>%CT9o1XdBk`A#"v`cM7x(v-k'B<gn5/A@](8*Yr $P4 \BukhHw-v#4 g~ nkP] x-:7i /AĹ:PH]~"âg\%r@Jbθq^ !<F?ϫe*9 d&eoZO ~L= Ur>v1re GŪ7>ᰭYk?fCfz1{ΰV#e(g$IZk+3\bk -(fDHJD &?P.݃'|ʫLh!>~I@R~W$=N|L~=DJ^"@pg ZG(=r3% BGzmR}[|8@TWH|@z2o#% Bś!e2UrKUFTH9o?N 9S b@9v8iuqOy`_F) HV"4Vto(;(rHBQ,I:I'&4Fl%LBt>ch ӡ}L:ڙ("g^t؃[`NB8u域nxƘ!Gm\}3WRQ># 7+ޣ@w̻u(6>!b@SN!w]g4%UAWYL _L '/E S#诟\f/)<pkq)!c`VӶK)]I;  1]W@B5?S?ԧQtMP G ,ۓ.\?I^Z7W'(p[>s1ygLgLDQwmI_!.;Rwۀ>lmlI>eaE$8~=CJ{8!pWU5]\rN29^w{f;q̇ΛrC¡}P->Ѻ?~QtR$&b.[E߼YEMB 7wJ)Tؤ0wyZͰ2րLLG("yoqr9jyg,'r,'Ŀk}@ vä@v2 H@bR=)9CTai6TܬQtu%A9ލʳ6fO|9'6^]b:_K RW֞16^@QHnqp8uEb2cC0w&9ƍ2!(U`4g*WNZ]mMgLX#@@vJ(!BذCc-CP#fRCxVv=( l8(Uё.N5d$F "`R9\|Rc1Y]nr@ɷR Z'l[ަuQSԩ,hn}~ʧh9(Xn1}nu1H:uQź0mwk4պ51C^1*ī%.G?~SDߵvdQaE~n7|Ad>((F/N"ۻ5d%_ɓ|Sf81[ޭB.t ċeq~NGdCT,%OVKԪC]4!Dbyx+J0=,PPPznέL&Vzp|cxx|Gy XyDŽ& ϖk:,{ֲ9@'6|~Scsޓ0I1о-.}]#Zi;d~2!b݀[<%,*:hDRgzųMXeћjMF( u*~?kS Y4NIbЌ"KQjew#3QP)NRbPˆ}.7:UNG%bS3Qݩg <V!w3i83UuSw;=JOVnirʐ@GVSˡփg JQRK@w}:׈;褛+Us9K\ZU QűA?ԩh/9iKx DڸlF_vڪe[mj)\"Ӈ"b5ټm r㜏~<^_(CѦ{:_ {P: ޗYm,8}j6űG?u\&jpyJcޚ׶"5sI@1$/(u nz_Ov>~ }ٞ] @esÜխ<0x{uNY8xCh%&IC^eHs 5eeY11$}/y fp0WEQ8_,|ԤHB Cf]|բ,F>ØJl9=Е0)ؚieD0aZ.mX#/w=sUpܵ*~sNq.lX-O8w<'~_| wn3?YkiBȍrC?\qaGty|IT*Q@Jqd/;A]%yG&'n 1~EJ\Wk (ej'"kbJk.z$QW!CE+ P[x{}NӤ<`kF;'ti~B)Mzv<; Y>&m^#.ٽeVg&;߬MۼyṕvDMtc""'Xp.2 *%%k=OajJDfiðKezK, *H| ?tN_B5_j)s- @5x5T^ c:2Q<ćVЬ*zӳF/( /Ƶd r6 % &a-%8(^|luz?vRqB`[wK@㵝>@/A0:+W(rYaEsT1uJ͌.S^,^. W<#zIHqrG3|͛Y$$>Bui @6r3iqг*s2%6,:v/3UbpumӜ N5 Iq^""Ri>b[&Qm."V~Y5GⶳZ|͢T6-"Z&]T}uA*-=qL,;d&՝+q0TԨt{(Y6.\[dzQrB% _:5'PkK"D)B*u19}nVQRCVQPqN+)!PW9ث܀N(TrG 5llGUSIt{Krm&D#dt߲8.+?NcM?$!qߔ5{wk.wƳ5{&jʓX+r(oz )M"ӱlL5=.8vZs8;mǃ:VNsQEtf}yVPd B!s{//@Iy#xO_ dg%^5DKjVN$AvHmR;cPiNP|6F|g,ZhFp^iBB, pQ8|i]cFew% g'1LiG,i+ !&aVe 3D>^6Sc+Nw&ŌK/u /R=|.4},~8/:è+ QM\~2NmIc\q3ߊCyOm}xszR_o֏X>cyM~sG1^zcsr8|0sV9ʉ=M+"Wg \0#Ok*!|֚SmZ7d[] RNgTn'ZNKrm[MnM̐W>ES,oLי,[;7ȥrJ$k4uA(VfyufIiG1mC2bqaBK)`0 SܝA<}mhvlmDkI8l*Y]H|E4ki ˠ08dDjɌJ]Mi8 ]mo7+?.v|*)k܇{}ł$Ed߯8#Kь̞ADNZzbYOUQ轶c9K\Ji[6*ۄ^T!?3ez'21Vri7oıik iGo& ,(TFmҵwu'ID aS,(BMYpr;bEH6:<hu&5E魚. ]{^&7)rK5GWޞ3qΨ%8|'VU0wN؛Zb4[L./Po{lJdvܼIOBYzrƃE6SچmYH+#`D afl!iTTn\J,(rKyIۇ%^Nؑz^K<6y!)_.=$we,oiJ9?lӖpƜM.VԂk@3g%B=^4x**JԢ`Lg?D~{(;92[]i'=4? j nY~CK<&Xqzζ'F 'n"HO>P48<94G踬=x|ш;8ҠƱ{q!OD=4v/QHgv.>fĶ 0rTMlR"5?H~-8KU,!*-b("]hZ֝StmVjiɠs@pV K/@R.sZhia.!E0Iv?LϻRa%+dOzT(6=+fa J"xI B=O޷NlVTCԺʉ֤e㡕2OާNyZ^?l׍|.ms$l,4 fi = Ş< 6{95=Uk0͏ 7cep*uG㔵tFD{ @9]CcűmLg^dzӉu߭5v6G[Qzaȴ͝O޽:m~]妇Y~^1ZBE͖<^JsGu#j部ܜVk‡'çu3l/rѱ|NcHӇ: Ԁ50dR-9SjJ$Eɔc%^7g=Uڱ_S:׉u'޹Sw#v<1V9]X 華[n wQ Y())D! - 66N@$O@+A>CDI~+WI*.Tr0Q,!u :E%'an&Fk5jo="͏v+uYSZ^ ꙣʥp3ύw3?ݳ t<hy.˽rozhc$ID-H~fhr<%)^#Y!Da95ѤMaripTiRH g f2-9#+>4N񳧠N ݠIʂ|#Hpyy{sk'wƜ؈nɜy5վA(Y551ёrws,#2ɮ'=܄v~b#^[G^ ZK"- iتatQ,n9E`[ d|0fwΙh ;3oF0V^. ?M{܄o`jm'SMw?r&4w\ɚ{2O洃Nq.P9[io &8 >P^0s9 =;Q &E(M/^N1侃YB!za8D)-󸅘bX*%@rp!#HA䄴$i0$^R+ e.ZSt%H4B2Vy0$/+Zm2FlԼUvMF#р=!wcdsdmS^rt,Z|7S!()4^3d[-|7V,"xQtF㻙{z G_M\Y `o"O>⧇dGV(+[Q! GEFXք\ņ)[Ŷ-k-+0^Ucn`Vu~qyoO,8;x<bMx.a1]/Ovq~ws6Vm}IWu>|{˷|z-|go?e7ҹ_ƫOY^õqX\v(Y{n~^kMy& ڤT4y;D^ ^rӂ'W^P05|'y;;|'_#q̛c@ ݈^c;M K5>M&~':q疂vhؽ}Ttag DmtL!|]xERK-vKj1 ڤ\ILU UgghJIRUnGRSx򨤱""r"u0b Fc"/5Jٞ>ľ&mdG7`cpH=_;$،Q0$Eqe u3ٌ?y' ˲9fhC=vNX_\]`V# +9ce=uelD6xIg6eߋ{7^n3YgGgVZ1 ̌݉ٽO=0u)3m!Ѽģn=C {`,Q:jI7Sd&nh%Wѧ'A֢Q:T&T_EdK=H Dv1`LaJUc;AA@3r pmiE"| cdHl=4FbuuY:EBVv 1۔Hс#hghrF .Rsd”XJ:u,b=̡(LQONpRk B8C(ڀ:d"L%94hF{VR$]kյlN6@} 92yY1zVh[t29 @" ;@j}S&EdV^ൗ#v ^IAc(fs= Pj9>ۺn :aGжve{v*0_mn?U$kJ5jT->4׸f^ʔj˗^3HNϸxɆsԽ4ҌfP &4 $OZNi {=Kwe<2eM+S4ǫx*iycCrKvǃvںEš=g/s?Vã$}UKC%TO0*ckj K&Ӏyur.jLZ*sM T-3Գhcx"?t.F(a AhAaU3g>@j™+01^K*+l h]K(Wn064{K!LcCNko$ A 0Zfv:PA Hr2?^=J?.VO530n ֍H%ILI:RœMl-ꡰ: i+T+wZyI L\-N"oy;9l΁N!(eU/{uިBm0{ظ%ܯtO냽w`1-*=q6,Z^ }H$5QT9wEe-)Z Ĕ(Y~0`0&&41 O.Z@KK$^f(G3E$a(\lHFPL P 5yp?@XTD͚N\B, l]Ut>R.dC)(¨ L:Kk/mhIiR2(u"XE)Pf5Z;A÷ÎٜԀҋX|@м-|0p!|=2Ņ"{ن):/ɈRz}#\A"Z bOP(-2j䭔h<2=M)@|Ս"%>ެ7=O/8V=ȻB'T$@Xp0-ǣ [Pr}SO߼߮ۻO7ɿ"mEYĻ  >aF+rKd[ hZvXYw?mRgc])Sƫl>l=#>+L|m>1`5e6Ayj8K<(Ya ޶$#627Bfžw GEۑ>MO͌nf?&{[p@rr?e#DűtHzr")f8O>j*4Yfj0JXkY`(ْ P}VطE O)\IՉJIdCSAvyyD3~Z\\,7ovyxC@`t]4Ԟx?Y%)Bs6:B`)T]GTOWxg> ?7FdӹP:Qhsayޅh 8O%jz [;,FȰݿx{٧ FG{2ZO O%rqӤ}9B:26zBe)/:qcYOM&ڷzrG~~vBO#B\g7Ny_*O"p.JjI1 Kљʱ7LQ 249SGo,15YTFsžHtVđK*kЍv@ *#ltNb <#g)Wbdg,FM zv%4si9VXa ԜQJ5j$Մ9(А>+LQ MnR۠ q*2 㞱 S O>ʶX۠ 4& "2s;}Vyhjm zw F ~.ڻ0򠷇ԏ=.Qs( m߂*e*;{y\B~o&,2^J\wCgǸ<60X lڊ = 2S}c}z"fl-0㻫]#/~6绘yts|ٔbSC%tH֓w ]rQ:_p oց(69gJ 7u?K%Wڛ!Fug ҟ^5|9[!,^{ S&. bԞw<^vaOvLHwegoƐ8oj3x 8Y|93}T%Ef!+u&BS@>pxQ R"LAfv] y2EjpTP*os )#tH'kC;'sL$_W "yS#6Z4Q(/XkrYHBm%CB_Q xqFfy> QiU-@u tb2i$+*8Kac@(H"I3jʵY.eCvm_xw2S3Eڲ^:PB$o; ":X lR,' qH^/K'\qe9j'rQ[&ֵKp^bSB>J^I%/dT7-̠SuAP鐉Xj5JL}1:&᭕jS/D[{rŌ׎:p7V!Sbʆ"1!Lj#BzN'\N´LàV1Ll}6"H9ݚRA$5+@\j#̛-uOu-.E;^ 2>*ӒvRqSL!DDvIk_tVSxr=y??=U܇QwSDKy>-+jqaϋ@d0-VJƒZqm=m-btXUZoF~}n_wӇr|S5_G)5"3qQ*/݂\%vSXG{(^[M%ah#4r;kq\/J 78c(xVZ$u$-"02ZJ G*9:)lŁ}H\'y+lY uzMmRm7|4@mC5qQi( PMAŢDB uhS@c;B냵~A8 ӼEi7RJ:,b1 %UN=9ZMMrU*$^⏦=] G>r=oŀfp;[*[̙iDwL)vևr~UV /AZ%;gWS8 $KwFK:+&%dO 9,iiháFc<x$=k$xE=:pݏ-9j+4v>r jۙ">P sNDm;t^/Kuzq_˿U{Ne4_LnzY#iPK7.~6}0++B"_%\>{%D2$[ڭ&i=6}Dg;h*O;-׎nSHW.dGqnk7z@6}Dg;h" u:/[SM!!_nTϔvq覞)l=Vؓ *E?䥘im4MISz6Q[Y(6_nBme IJ?K)iH-w bV|G[l[,jUQU6Al,cǨOGf&;avW*:g tO^>?ks7mNCd3A{ KtYr8 ;XUлSr(tc> ۹Lb/t98'ƴh?}sPqړq`Ma6%+S7띉t?yh"hlq蛬Q a;y7C+Q=w^]}z1jϚE @?Akgk%(Ƃ}ۤø3*+$,PސetWl̃/tglޓ^n kvp@1CcN/LIo77 UgzI+.e gs&\-pYUeW]V_/P )Ndh,K cJ@\6=żgޅ'y%Ȼx= \߿/;ytH!:Y>s3^ǘw1b}Ƶ"!B$xiD̂D.u"P`~ 퉝A!/D#챦?eziP/U4! YpgXݱEƞ fN׻/۠Pz&S.A!S,)h40P^-2ʞd350]g|\\͓-;>5>glwV=т( !@_Q|4T3M M8**5nYp'-+eYy{N`@Y` ULDBe)y` LDL0V1dVAQR@9B#ѩ>5FΐEi{ K@c116 V7͑.fNJac-v`c=#f,,wMIm-c?I0fkhɻsQWnn"ԔɵL ꗛPqO)u$@bw8F"7~Q>h\Z< ^FJ<\{i3gR%?{ɍA_c\d$w6noZzyfd^=yHd3+ vb_U)b d24:hXh䃳͞w9cb95/^9",b۬iYڑ9jt @,\T{F8hA①oL2u09P(١A^ah_O yRHc+Bk+$B8k\ŌNffr0P|/O9q+2mxa! px0$0Ci Ԏܡ@*6D`D=.xOV\Bֱ>j6n#G 1%ӛōLo"m$zYDl9 /YL82g+f(jaQx8 s4~cc4c|1)o G6Fc l~\ ՃL^ѩR:BHq  %&*R MWwA-qn=Ш*e,URD2$bjcaou7I$H$vCo؊|!!zЛBB\cBa~LleЛh@2Ƥr*#gb@)g.f%H`TXO 3L7wڱbEcL0姝 eb4dgj:`oK&V#,[ud!?ONjlעѼZ2oi]]!݁i #N-3 e< ǬG,[&-eMyd&x$MgO]?=*&ɲ,Dt#UNOM-чv MTCDvG Xc 7rmhsm Hk:v"WPwbގ}D} y([]SMqSmYvՔ16jQ 4wzg'VwpNd҃?lRQfod<TdZUƛNV9dN'CC^V%勺[lp+y5خELobuً=h+W*b#mAͮAԾĺMyh֭euCC^v)%.ryWuYaO9YRuo_1 XYWD{шU!o"xuȬc2dIzӟ Lq5me^mg'#=7Qς?u(VC@:q:z>.K6}<9Saj q"-#FyN!UCygc ޵uAV9D6Vp`~#©;``#!ˤP@ݍv6HwRMF/ 2r;;O'n}aRG J0i'JG,(e, L DLJrd7R+~ Pgc tMƊY;{%_A5ַshQ=\:S &/ Sw\%Jyr%]qS$`u0@ʵܶnL(AE%"RۤJ0r`!\E;}H99r ;Ay6C0ƽ\լz0~)MG qkЅ˥z eT\Sݖاd e򍹾^.ݍ0s^Ab*?f$R"%QdB4J0Vy%fiyfz)( Lv;9bI=KW C+'FGS<1*{B˘#i5souʁm $3$gy.̀6Nq {ô IA4F1w2>h:ʭsq[UCv>1Evr*UbRᶃ[7@n?.E2Zy'=Q*VoPd*%Or=R<w?~F-gW6C,Ѩ=3UBaR(#oaHZ=I_Jk.وBKQO-;*w>UrShoۂU0g*LK}@jp3up__% S [!F+"I=lA@ eiK0C*BYmp"7VW#d] Z9y],!sZÅwW%9.qwii fWwmRkf͜~sR\+Źoq u >/~nYsAٙ"(:(n79GbjBbپW߾ jic6r:e;q%]vί/|\߷ 2h'DODic0! ҒX饣XlTp/+,A[JbAǓf twr13Ds},KYX(o }*S@fK#ܛJ%7K;p)a=@1*k 3nvބ:Pi{B?B}s ,5= ]N"4BT;S3H?a^GzD bSk ~ IKCp V֥=l0^Y$jé@A[j&J 71F, :˄zA*RB BӸ1V\b'& %w`n}Ğ6S_~J_#@)cYwO۵{q>/x3 ??ū]5X-ܯ~4p&>0onɧ~wz坜~]D;~QLo\D;F\w5ϮngQӓ;*6c>4gs=1C )'CݵL7 ٺ E.9S\+Uб:1O(=71r{hC: JJd@e=^2 ̌@R,`?믟UI#ZͿc <85 E:D3,i|玜t(ټ@7/w5mGy2 D 0| {63{v`[-ϔ9?lR-1y<-%tF%Cuјt[(z]A Pý\ mFb˻w7\0mC#Vs}4 S95Jt7Z9 o:=Q x<+K,{pӁ2Z ,yq (f^l "Ih7/oVeVڶY r`ěʔ KzpnɄI2\8G6. Y5kBH=̓"-e`H _EaXmQ;eQsQVqԋ:"% ၊!JorwA|SA~Ukx`S|_>#.9Ed "[$WE餗)R=+ǚHD@ ICU>I>q5(=tg#5` A`,Zg;ٖ !*>Xw<{c!SRLuǒB$Epd8FV5 N g+X,<5F&*3L2#L<(e B~/OS"DY1@{O)*ˑSh&M @yO񜄨oFrS-aԑj@h3kGAcEuL:uAJExҦ'd5N"OW)y 2lur$cY$]]Aysooֻqx<|_V&Utt:`ǃMi/WopCd9fb0:@ƅc2n$'&幩IvYd&=ave@(oHhwk=$tRMF/CB$о gWύ.M"9pUуQ z9j4(s^_z9,1!`#K,`DscG @ThIK(PHrNvLTCS((4iz@d+f iW3hǛj8|w0=:`>+ū',\%׃_.؊lf~c(%JO|j4 lo/zbU cZ~YE;gMKkXUdОNwʥ=NӞܲnZZN;J۔B6nÙ֭ y*YRdm3جgf̜[M:/y?lR:#c]s3D wݮӍZJϛrEDJz>x2nAbC, y(kC#WA*d;3D1 F%?s- )I[ji?0*{:eۍGv͋Y}ׄ:(OWCw-=rH,ج%2|5t5f/; ̇FU4%RVuU7jlȖTʊ `SFIq a9|?9FDIq+fb B,+vL+-0dq3g&c49\t#XTZM3B慓KNm%;2qGh:]:T2^\ AO)Iw'VFEnk][amtm%M+5ys0OK F*ߦ$JZW 44-7s0VkcjGI.._YI74> $j(f<ڎ5xξaӎsRj+YN:.H?/l5 ϛP\>xd[.5 |v[^yݫ39)YeÍ*}<>*ۉ1Qm ܈r='!aASwZFք3I;c0K+V9=0Vf%74MN\4)4N&DFYW{:q5v]cHM*ht݅VjՀ>0QSt.ONeurS%;?lk{NX(z^Bin5oo7,% wO:8gA+MӗkoZ}Vߦ~gVU÷#1}DR!`C{s]bӯom\Q7>l,I#X9;$FAқ $Ux I|W|o~39ߓ!J씰^4_Z: Qz<}ޚfEn0is34pqńx~E15 6ݻC՛߶3cH.?Qߵ wN Q1&CcFn>uEzF #o۸n:jホ5̺<*wnm盯 Hpr̓\Xsĸey/6PMMI(5oe. wsXi#rHw,`O| A9C51tנQ:uf_  ^^}Y5i-ge<Ȗ rY%ĘZh<VmZj N xf<sV ekTy>aX#J'LZECBds2G՞'(4YuaHi"Mm>KlC{'zv!Xe)v47Tf\T0aYLveq2ҙ ]E:ujDJ7 :+Cu݈4ۦBU`*hcKC1c/GGЖEؙhF <?ZhOYfiR˾{V|2zեΆQʮtpZ!wfo.M[F39mo!ТFZY%̄ԙA_=ҌmA[k=kum9Y $xSЦq2kڧ5+UZ[]ՊTKqhihtݡ'4Q-y[+XX3'iNHHr"gd\DK.)D$hr*Qr9H ^`rtz^ ql{ZsHBL JU͡u~e0s"ZY&G/J,5M ÈDظ6\>,3* ~}g<ne 9vФv^B<6ƮҊ٬eh~YgH*˱.1u#+/x|e}8uc"e 8|c6r+BP 君S[,',(s\tiN+Y6mi,4u-Gk7F9{1|Ő8i}hWM0NX[Y9kY k5ev W;_remŌ&.>Lz"t\{[ g!/Q`t:uҹdg j}nCr3BjhkŨmu48i2 Ɔy.ĄZ#芡fnk@RVN u| }镦k wq#ֵ!e{"1 hK;4tM#!6ڿKC1~#hL:_'>eJ6 < ]av޽? Kg +F99{,?mQ1r4J9c0MGHZ_W*Boi9O8J/T>?vb;w04Bl7&FWn?|N(mڦ(FU!͡Ԯ!. &oI=,bUoN*UᄡכS4 !i-nmsm_|/w6_۽#ۛ˸u?]\&3K6eiߦ,۔}>>|19xlRv}/?W[w)5rm|wstcfW! To?΢.X_lͮ&+G'O)5~ڹ2FYpދHf;VzD XTQ1z Ȼ/cWv}a+gZJ9Ca M2/f>؂"/xlAܾ0;1& Samh!cј0A;o_RtW0!7e_"@`suE"\(iQl'ٕwEYtc84,Z<,,LCJMc_Fڈb\_޸Ww 0]=]> i Z]Ϛ {b[ '/".>: Wg[H,F8Áp&gE`)O?V#mYΥĘ"PY|CXڹ)EZGBLY/* V@" +/ҋSt jg fNOX4ș8 zߩ3e6^?9 Gk7F9^hj̇{Gʀ ].=㝩o{N7OfPە(\7w `u*W]64ٹK_dGв*"IؿH^bWB(q]\iԑqBۀ6OBTR.!]]dx<ؐh"DUDux\)[{XMʈ9<܎Dž?Qp6sf(B&E$c'N2rL^_4L |"RWVxa0["<*Xk2`G -b31&GǫD34F:gVL.;>ۛx_41dskcGE8/,2cŋ}X+A9ݗ}Ȓv$`Xͅ&M]HU_^VT ]kVbyAʍgA3qN,'"զ2vm B]~,Z2xORg5rH&oVFӀ=_rV\!U:Z4xtC!Q r$t~%1(򌎮*jܱd\yt\pQT1nbgI0 OdJAFeR~eF7ΥKnkc2`2M][#ӑcuitȷΙ#JSτA5+fiZU{NΤaϪGnɭ1#/z_=xo\ x@.UԴ3mմz)$$VH.Zs+ٙ&/=fi;OTVM+d츂~wCT5৕aewmSY=\/,ټZ4*3fe5D-a7˟~z\,ޔ֟~޹ۧ^Y2"t/{?v<~Y^tӕYzGRVek^ew'힠]ƥtyi0#hi,tۖdU+y V yͫ^-s+)-*u)IlZ,u%uU]G[ ʂ1[9q I*M.ǭ;EͻKSڪ\}%qh"%qtBIj! .ÈωadԵ$w1}8g+a%qegT;^'{O,ݕFw&`cMq\W_wDm58UĝoE\#:Z)vWwR}l )vWy l[k!wNSQX?M~װ h"?>(Q|TNPUz8@B)Q2(r FՒ2h0v3NAU*n) D}.G}8.9g bj߷Y>Y"|;4q=/_);y=6Ob#PH8O@ʥ2suGah0vܥ2uU5;K4("cMs>9P$ =b?]w7tJҐN٧zt%vtE=".t%Xsr0_> t<&G+Cӵr.k y\`Mta<u&oa8"n0EsęKt#zq<'G9}wxpJb)Dӿ t J4r[LT?P);0xOv,etʋ9LCSL@D{T X<5e' gB֌OwmLUVP!]\PqeEi*5SmiY>]jQ %UJM B +SqI\iV3((mZ7udžu;GwI OmX0.d d յ&m5Rh- &k;R[{9gbJcNfOϩUJ"Vĕ x kxղRmSI#* lYcFhWm b1$|nXUph S7F6%5Mle]ȆB7hX?ɪՇNFI@ׇ c CčFힼ&+ȮN}J0 ëR9hi.N}rwO3YwK9*@-::xA,ԊԈsU˸fz[j%Q< il`folOoliDU5'ōI2;́gl1&-0-pg  Ep>nOSrԫ:^Ro!56Kx֞5*2pd O?礉,Nfyվd}Oye\}#F%FRJ'ݘ*%=Cf'尀t%x!fՆN>LUzoOb%LJZ{5 i~owwے 6盻կZ_o?~'ɡ> !t]C Z^]~l /9{>8&= v^do\CѓVpGA?(Mo [1њb2k),}`3c'/8,䨔12]wx5~<=޻HsRD9y0ĹÿJaqRcB=/1{O#HjDӀ9!O$GǧG ,dEocyHÎ6k3]!nVsS]ݷW>47UK!5$Ѹ9QϡQG*h/9<Ϩ4gFfL\r')KcT=˜QuV<=} KK%(SowjǓ]ݟt4H2A<'!0I{Ln݈>}mx;n~>6o>u M~{Y$.rNլ78$~shVmoZQڕx=|0WNaq4\Wet~ӻmqtuc&M0w)?Vrr\ gMy,=ͳ}c[ $?T~^ҿxpz4/wsr`a<="r7 *rӉE ct^=Zou( ;/؄[H @$&Wt(8]Og;"ACnr2cEm(Sر oj8 ˩ B =&?q (:OqGdO`x)@᭶q\m?^ 4^K1kӝqSq5{ϮS]n4ۛyY<>5i7O~SgF~C$QŌ%55%ǢԪ(yk DfsMm2̾cyn<`gjՉ72~UJ0aBOrQw׬iiͫ/Mm7l\! NG̥Peng?N\ۅ⍩5 Jg %+MYc]4 YnUK 4!3mt=yfŒI|OnLs><[5a#kk56ϵJ0^55rn15 KDHh}jΤȅ^sfTﮗ* na*~%1bjKzmI ݬpN^2ݘVzHmQx>&+=?+L=l,MVzXm5y[)gnVʗ{'J98J ]DmAPr85+v(=_WbIr5*nt EY*J {TdX)yh ȬLK@iL^{},PYDkuuLp)T&ϦG]ntc a}wa.ȅ{7G|9쑛CPJWL"^Kʼn$n.$pj B97X=baC c/ * VX'r)w:,~{ʤgPj@k1[{D- a\Ӂ옺gNzD a8c\GCq~TL4G7PoBK _q2 ܇Gp=`:rEh(aC^`V\Gq>L*/AFJ+/m8qԼb1MhӞjn\F_ ,:=y[EL]`,;5%HqNc Ufd-yάgxEjt޻jtgg[EcBVC+Ǿt٘/7";.I<$uCOe^5H[^VpDrhp|72*E=s:OKy.PkJD"nPܪK.d={kbs{`5[Tvם q!J7%`v|>#ɚp]k`NgFrb3Kz1r{p:-fP9M2 7_޵ٯxu`a0X1`ߝT*+7<4 !w;5J&UM9pe}KXp0ѱ8] xd2{RV(jkF"͚}D.B g&C]4~;X4~[ 4~T2y|[:B2yIۃ[үIT:ma[Ln<ݍFYNp)FFMw==V^z-$^S𬿰3Y.\6b A7VNl-M~*Uds9vOAvyLpW9 uR4z n,)2g";xɫlˉebL?5U.µs\o 2 zcO. Dm{OZ ]Z&|7&<$B ڲZ}}fc5i9oAj0$vJϗ*{ AdG`TXa OYiVl' Ty1xlm"&R$aVN5lRk5bQ%U*I8QIpʭqd⯻}a1UdKu<8؄33 _8[w+J%qSk }x۠غ/W+Or🊱3 b%B }k>Ù}B +eڛr.z[/$Uf1;#{|DMGv4}3c{\ @#*QP(ުAXoٍ^M&XC׻EDH:w}&9\`~ lf~ <3K텛 oM(H m޺Ip4ACWȼ }13[>/{ېB|z-:VN:1 An0mwqgs dj&*S;--&ѩT7j!02,qgV@22 S/* Hym Di iIE[K^QУ>|ôi yu0-{9xbg|NŮh99C *q*v}㡇2ScsjaxBۉ7WGu!ݼ 'xH܁zvN*U'h|fdOϮ7WJd^E%ъ^ 䈟7Vi2_n* x5>VgJGOMH5KUZEm+/Q"˻e::e-F9"XՑnMujՑ"#UI  o 혫,+-iE8m(>|c E4Eo?~5 2x\ Ea2'5j9=gnέRfH b@U6H`r}E]6# ջw!<ܶ`N BLb ՚@" АL9`-"B %,5w;(A^N8Qq=8B>e^6_ Q*53UkT KE42e/ JEDa+amqpN. ng" "xS{#+wrz5>=ї.z /i]eF <4Y]Q`+ǜ]W[YcKt4O|&::]$k(љۼ:6OsOhUMkRVc;@l]lxVϷo<&ftHT#X^E_W+ ^X)ؑΖՄu\M?Oslq*eV(x$1OiKN?c'b1~:7r@5O3*5/ãRU61%ܐYZFrK[TS_>Q2-εF?:V?EUx ꤶ*l zA$oLf܄*5])\2}LY.;V~?_R2k&ЇA+8מq8b6iRfDC3&AbAX+p Pr%bӅ%cmUf/{(=wA&a,{yn.G&"gkQ0aV9Yӣ[X.va8V%#]`aV'H8~wq71p j⻝ gij'PTfԨA UڤjҌ&M$\s*` W*~x!fɥ$+ܰ," Hp%P0 `G&x +U]U5v49+ ʕRW/۪u6}c׻V%Ō?p{TDee?W~3b!W!@RߛEUQP$GT={U/1*s:rH)W??t2@0ÐoF{Fn(6ϝVzR|6Qf ^Jp-'T T[j 1p\UWZZ=y~p9vG׬*$ ڶZF0)ȵXEp Pq$HmީP F !9Gg۝knrXFOx jeZ4s>Y$ ه[V,f*[-_ 8`[Q+S^.35,gɭhxh\ѣQ]OդvJV|˅+o^Yo2FJ^\NL*UTw߹)sd=)j63)Ŕ uY2fR~VO^ugXN,+f1?.}(L{]EP2&L@!L8CIL$ZgZ4m&bXH*Ev͜hM=1CCml?H"n(F[03f&jjdT%&IC KzF #BfZ~|5ւz {b FX* XfYƕ faCXa.0re(!(1V0<6 >lhH CR -?p f\B2 _\ڟbN0b~qAkBvzedzz|Rv53O VEL ZąJ"nRQ^ sS.%_MחR7{ #D1aFih9VV c ő)Oo=ؒz5$642dSC|H2}={ۊ׌/ߍ_ۚ`ix]+ ;b: xo.=ujvQˋ{p%>X/m[;psZq7\sGtZZJ._gZ~[רq@"27LZҳY+d=wW {Nw]}i|O˟q/ ?j-%!o^=\Sy\Um:eC\GѧnkTNk$ڑGmKBxctU 5qe7'RO/=ECH$Όhnw TԄ7a `JM +Ej[VK{z!55veSc$ ʥZ7^kb^7>gФ׌kkת GBoݭ?~ET6]%z-ѦV*ƕ,o۪Cff>gKtNXnD[*w65 C)\Aw=wsfQȪD3㴒=** BKLL{KF1e'/v/:P9{ʩvW,hf;`ߩK*u/˩Ѥ[h- ySɜBt^ߗn(ݚb3ut8A(tk~S!|)ϳ)cYҟKsM'Fٳk:Z5칰IͣQ3?G_|4'իG8ȑV<^\VZ%V筴 soZf"?.A$Z MK^wƩB ^{I،/eŤhC$dE`F 7A.II x6 9~q+H:=aW>~.bO5}yBqؗ'֫3~ Wcߓe7Į(& #%gS{3.A .u/`w]]!ux=ԓJX% dK vibw^НN"rGh.9c V*JW.seeӢ.`{.UX"r)u vs *A\aDwpجs%6>R!6Qg+eտ5X*8B nsm -3T*JMbZD  ƞ&z o% |#o 꺶Ag\]*H0)ROGy=ԞLlv1ASBxmr":sWGIB,isT `$u-W@i=*`će(yS7RD)$̨Q7)R I90M211HThR w#U?;ToBo.WSZl.-v15]QL^u!ؑLu#"܎aq;`3$ad+˙,_Q֛E6Md&b=Ū"Urݻ8-X_^?D]zQE{ɦv#}`=͢lGC؆c\'$ ӯIJuaǓ[;<($Ij:>OX)hn}]'_ b B՞˾Ӈw?Xʷ5ՆoyINW~/b.q dR[nr$! . `Z:̥qAT=x$UPp"iXAY!D^=|5mߗcU0':-9:|s HwrO;pof)nqCd ֈ*{N aUo2jq.BZA.ډOuz9ח:tA2PX!`=.slV]¶0) [m)%!_sR\ﹺ M#K0>Ϋutk8ϫaLvgyOI8Ex+pbƳ xzw 7oD&# $w1dW^|q<|v>L,?]\fX ϣ#}6J7߲}Cջ09 zNtrhh;8%+*Dk'EjG9AD]h}䛋+Kx.GG-% 'Iy,O>X3 )D!4ʜPŠ3SPBp!5GYKP\B@PNVve R)6jlNJK+GRkVz*V*$C˂ 25_X7^mnH(ح"t ZZN*׶HӧI^)"ذ2B % svtͣƭ;Nnk' ӏ1-A DI\+}kHį'`%=CH!V+X6u/k_Es8Cu"PѲ$1ynCbȨCl(-]!E1 || ׄBUkSM׃`:ߎ-G]A?)(K8EMM, W8kJuÔsҔ-DN5P{4Ϣ2 ji9x^e!#K>aUfJw~+)#,!˙p04]j"_=uBrz:iunryES.F)T굕BtR9NfZe&WF#MrqN=SG1Nd)k+:S(tnOtSFGhFTUؐVRbC!iʆ 18 0e-И#ˤD)0B(3Dr2׻sj!1mԆ=%+3Ɋ4Ǩ-"T:gZX~QU ˄@ Ц(48'YEE9"B}](W~a:Qrf!!Ĩԥ  LM1JTc!Bhhr.LqiK=gc QI[qLr^W(9cɴ,?O(y)2 0-Lq 2j}: İl=O@CqQ?wC9EB0GM$,FTq9͋ܗÁX!$'y!C22*4KUΰjk@5+fzc5u:oD]1Ue0xyHcMh3|+VvYfSk!zW̷OLGahS~^ 2. gn>$7/mP1*5ɇoD14 GP-_Talgff}gb~GۯEW.^s8^O>5U+f\*]ˁܔVr J|`zr.rĝ^CF,4f+4160M <T}Od"/nEtX+lT YnZ2uU$n CD0;7=-V&MKuŊT=T-n%r2ixfnfds;[- =ϒfCRan4377]~hap=.gO%<'SЂ㡫;恱,";2H>_|NR%[C&[Ǽk?ݺ)%~JfiJ|?FFϵ%*9?[4>7殼rtrZOS ĩ(P^;8JLWa5bUKe2O;%z5"p5'".5de"ąK V?{qUy/&kUZ#4#ܒZv Y ~Ӻ̑f:Lx:~~<_eMb@f|ۇjU971ntF;7~-Ȇ  #=Ive_%;an^?yl_y HMr&dS>Noy7…wKA餾#ƻu{y'5~U9лa!'nI6%P3-FqO-U!*{4Kz:,MtætM\-frW}|ޜNlnp;7Z}pOL"i]w˛W_٫8z+'jW}*RjkNE˹^K_m|8F#UL;T!T _N2EXXZʵj).V|FAB>~U&ܯ[Gq5qb*ֳjņ_qBE"%5>Q rf*Dl5lL9!!QgÒםpq*ۮ*\3'R\\k`#Q}PohU B|}եXJj )$!5" 6RWZ~^nHM.U^$5U!T9IΫzLl$5= lg'3ʔϠoQ>T9C-R΢A ̏4FCK|Z9`= *#=H)f:pQJPݕ'l0{֝n1jbtG * mVݭ;?.XsWyL!0R!!B)]qO_t#3ǩZq!g_݆8NcW`_B60"MFV0+,x; 3i(*>q.+V2n.˝+,4'X{iז:$C9 UDu!,fZU&Lha0Ɖ8ebgέ@ *hȖVsrSjLQ3UA%U3 wU*w5XȉhMQ|/^TKAP wDU.LE{WBNDlrGmK;e/ 8rgK7^Nh8{9BND7l*,$5<%nr^'a/S]yoƒ*˳adsa <6]!)g|ndYjE{8 f$IVׯڭXH#"bQ}N:"8 8jǸmKXR:*'ThV!QMTCZ)}^R*M֢)N:VJ"&峒l"&j ݷR,T:J6M]jT4QjMKe]*8 K.֏ % %*x@ q! yP7Rq{Smv w</ ;eR2-}L6 'n&P]'ް;fkQ*NypY~](nŦ:)V=R>_RBX8 ؉wQ A];.(d5_ؗ<2n=b5p@uP*^pNK;\QtN[R YjScJDѪ'QzSnf9vtAc3^m I4-jnqvOvTwRT?Bg68l8A`#H hIS;ۈ]* U?ѡC5n(5ϲO!RYa?srʋԺyuTpT9f:*:SQ%UE)̬ / yqQ%aH @ %H[2ʏ8`2|Ida`BQABUBBW]AK])˕ GQ EBRs6@RЋOI OIJ'C3TR XDs`d*U^?(UL^/UBRqW3yB},L^*a37^ +'xd8 ɵ*erh 'JER$z[](ӣ'ET0%PYb;to>O(-ܩҽ9M||yFU{y Kt Xn*`&~\؜T7yZwe}?{7le.6w"x֤髧~t5{IZn<6ik$@%IߛRD4PEn8PIv'ᄍJJ (z?~ؚ#;H+Ѝ)ȿ$m9+_Rz  0 EHteeT~@=%?~<y?R]kSO5RDγ*X1z+O&CTnuxUǨփ/Mb؈ϻJ9:my7gY'Ux䷁ɝ6o5!8aη}s$O8v'Uq}ȜԴ\VFx j2n,r\> =vVA3ֶBߒ?bL;mCRCjNtvw| `beN=m0zUqIq'ƩB#U%d }D%A"y$N`;KCi C1AXmu\0GBZ+ 5k>ڱĀ-TS:^¼o=L]{1eK& :!yM};Y`|IP&bPZ$DJ+Ꮙ\Vo-  A1{-}f;TΦY~{0j`q,K&Cog)ZX$ {w8P}n].z ! .<0|C'Bz+q5cC[W+,v,{h{uRBMA H#t7T BYJahBǬ %dS}FQYD+ZJ!vRhʯcA#b7)ͩK_sd066DNDc0>3I`>}sH.;W(PPaVIGoR3cdvg/z$,qr{9G i :\7#@8M Мp)nOO~Ob{XDs1ܷ,Y*qS_3g;xҝlO[ts+:xt~5.UrgM uߎ\\\ft}is'Iquf?ׯ^<4/__su_- ѪU^f4xG U綕9d|j;?OVxZ_|hS4I{xiB;E[AϽROB_6@Sw&])G%#o)sYzփle-I`< ރ7\_vohٳcxIjj^VI-^v')MSP5Ϟ/=xcD'hzd<_yXx,GH9HP3eAD|c0Wɫd8NV4`?Z;U?{|YqEQw*Z ^Jy禋۴o+9ۼgv理Pm7_7|xM"y$R#̿1jC{:)4{Z[Y;KlpfguXe_}cQ,f~ KO4|\Z-RT"um7]0\|)#q(AfG8 ~5`sZLus"Ip"&jN [wstǺEv=jՍ"])dpC@(DPR'D8(&r I$E"{i@i}2Oc8d\o|[=::լu'wUҖA_?'8-~A7ۮ@&KCc͏})N.ϐ2uGOt>~S BZ, / =, "7/Ն(N~^>D >ܓ,~ئiPf~y<#@*@zCPе'$l"z\61w[6-Ղ6ny/ZqRB!ۨYZ0*wYJQ)5 "aLT߬Sy̥r7)Ʃ!)MJ--I)k'R ]񟷔&DSj;4BJBL$/2ΜZ8Q CNd̃ a~2b"f&Q>b~ѡ} ]jݸ@#lYYGJra6,Eestb'|Kc'MK3ꌧd]2=Oe٤qMBNy~'A' &v7hl!T&*fC6z-/؛hw0W#2gs`nY? ngľStFhVX`$d 1W+qaQb+/ARGz"%hEOǬg3I<^+=ɮB2g!G$`S2o0ybp=@Tt_~v?{E'xKK2/x ?hʞؠ[ش#CPyvmŠqhǀ6ݏ]]4#&fzt>dLOmcGzq(E5^bZ^CJy BX %sU2Q#%ԓ(S^=Y_i@t.pin~-C$#Op &05te#"L%%2`2$#j24Q(}s9֤2IO eMV"hC ЃIɁ txPPDmF Ba޲z݋җ0V-Fܑǡ/d4*G}`L"ảh=™0(c^G2lČ%Vc7wIgGjrcz=$ 2 tMҞݳU@N74g.A@Jz[neR(WKܓ\P@8Bn9HCG#= $LPGn\:rn|&^ l@CfR_B!W[5/+@?x<`uYе2[/w-}wr-.O0fG^me #=g0 e%eVI+-C=X٢7oC㞠ZN<XVZ DDk\w9QgDN=+'ttOvlB*7n]}  w\ͬX@j67GDcHZiJrg@+}P^r=ky"iNf*~%d3:T9_T\}wս6ןy'q7%Bs|lMDnsQL6D\Q>?zQ\'q69[B;6h7Ly5oS7cbN"̩e}>*xy&Ԯ~!b)OT|j[i?B36:uO/`ꝝxx<:2a/b^9j?j?dn݃ OX݃^$x" dͯ#(RQ6pNyG:ՌR^[ȋ<򮇛2~0ܓ zUftB6be14Xt/n]x+w("1s; S-^6*Ž 1\t:W5BO9*jk}E;(=7'׷7m&,Ga*q6q箰d ƷTgS*'~zbqry: ېfFdƽR%vR(PTRoZEE܁(K4edT}\WqĥPZ͔~& #f)J̤Z"Bo<+4YRU eIyAa  f+&E_cD1p(6t^cȵVScl|; uL [{f;AC).ҕ殊1;w՞9LrSnZ!ˤHL5*$(8IF"hA)9fHYbRYjC+n7_|LQIx@p.A+*XYFLX$gdVd`M))%$IшD,'O391eq)"Ϭ %%9M%cY)ihD'a[f$O3Ҝ Fъ7  #ARҳe+zkm_KˬZc!2]WQeA *{(-O]f?okI PA H ~֖zT;}J #WW/=o_Hs*aX?7WjJ ԀZ 1-L1Glf@[5A+6ducXHn^VhU9콺{W7btTdC"=ԮP۲XfM5쌱 AD {߽͍ZjL>wZچ\̊߭o]+3/fŦ~=䕻hOiݤ䓈n6 FՅ.i ќþW:W1>ةfэq4^c7rWԀ[t+ܭy.ާRt\\!vBwKy,hǧ߮hsyYOjQ!n.Mr8ٰ~sWa~QI[q=#F^ƞ첸|*]EO#Y,?^].YGRa $O1T[?J%vҪˑ"E,(G)Z(H*xq^}t et VJg\'4fD # *2dY=qb$7 ngUWkبuH]\=uYwZܞ_ݞ[lT}RWd`wg FY^0Cg ߨ5P޲b3--Pv^5?PP*+Ai]0M㕙ňJO = {}C!@'! YOh(dhw%F-K+HR0gyP-L(cD@DH& W*9 rqc Z5DiRsIJʱz)%K}2OK) 272!caGU 0D Flv6bLh­1ePS eR&$I^BV$R.ĵ,9Im$s`[VԞVs$+Q)Xde sV 2ں2GdKUM4ɸshս^zYXk2&K8 ^vIsX3- #&u]8k>aj]{9-\4\%1ّ7+DNh Scq)V\ivU++r=ue' 7'77_oV$^r}m+ܓwⓋ|W &b<ޜ\$^Oeϼ\jRu Pŀ]5mՕ}Sl5 $]n/GͻF_Cw'^U/UwW ZEGZ?vc0tާc@ +"PB 5`t!EH߀UTskL z\q+O>ȯ$$U˾G(Q)h[U8AjE\f8~՚:%9}4zzpwI0D q4_dPs@~.AUC7ma > 4!ᛰ0\xH ^SJA"3?FA:2rOHctP _h"`#oN+WVA66BF]wvZ4W4KM+\S{OR0CkoЎO !]c$'*`NY7\kB괇hF2&]Ps)>``]/!XӵYVk~؇!uu0hxVn0r&M̱ʹ2WEwznzo1|=crexo&zgeռGF7gbLb޶Ԛ/Zsy=ݶ3+m>58lՓc_u/WOn1 D. #A ™ek4@ wG24$%c\%]\@ 2=`:th4ΙZK|#?İ^Czv/o'ՁaUB;'=[,Ԋ֪N^Pz( *Y[1XQ0VR#n_Pz(e: K$@)a(D_Pz($ Y0_*HJ+Q JsP~FfXs`]{WwnJ 3mc=?Ϩ>+ńjP kZ_ A,ߖҶIw,f ac|2G1VٮZYq/c* [3/?BW) [-*޶;O̜K[<>x83M7JVP#_}V婢ԏU6O=dn6 o:䵰?هy2W}n uMl9[PIMTɺw|ǀ灶\̶*^?$֢zmʼnUyɛ5Hz[k'蒯>ե]9be)aU*}&i/VApճZ@M-g{Vp| ]NqM-n:{B3NSٓ+2 ̺F#\-҉{ͬ%)BXuǪ7m[v~8̹0sC'Vd@.jT2x)3"D!dY.ԠBIN6Z4-OS N˂Q QJU!If- _[.>천Opx?}^mFVr&:eߐ~5XEh_r!D}@亻CwGWR@i XEȋ\f" ifb"ED6=򎧪^ǴX#yz^7W՛\9y e,ʯ| 3߆LάهS{0y-;6\̎j {L `OE/3aS۠|ۃ^TyXG/c >q:vD(d7e%C`˾.RӜF-)ؼs`"59Zd2 Vd/D4Q.71Ny^-D?t(,u_)#(s!TҌ6S$0`4W) fS"hʭQ{A ky,:L'[,Ӑ +[]O%[I鮜ӟ0{_g;bBDamT,4EA-"EnHTJ[)a2P2Dmߠg㾷 ov#Cw%oc4Vd:6wH:D* Y a@Ӝ9u”9EAJ]lf]ܸOcLtkwGz[O6E*vS?y O$>em>d׈oWw}Y앃KlP-T~o,TN4ԥф̅=CQڒ_t go>{)rwk $ zRv`'FK~=f[!0Pp4ϘyӃփ||i?i+ɱ g<_[gN+g\]!@WWfw55 o%T1dVY #w;ԌwW^h)Fv lPj+ەh G^j{xOGp:#=nU>( djlpժZmwsuj]s=ϯ~S_Z⩭Յ>jí{a1^V:(PIk1vfm׻wD/!:/kpդzcxT!^)-sg=Q-wzT-rQR׊wrl HFMhCE45ey<"͟n^^e}#R+2F'YڹVۤ#ݕZ! +'8&X}[?/m*F"ẋN7/! 2sv8]\UyFBXx昝ÀnDY"+}{PM'h&gm~'>cV, !t%>xd˚)Wh!C4 SB~&ݓj-V!Љ}G6\tЛw n] C4 ST>n(bHMh p6o:vuwL@L\(W?W۹5=r=NQԨ)?\},?g?-wt(7:/wkmowP+atɵ*SZBk\ȶzľYt-,ZҼ L"7ngIj#Gt'A&u@y]L"WR Gdwg[ZCJŦB NMb B?LgDqL,ȐNqx!G8 CQ9oէ1RE(tZ:Wkj}С_xߩ:Zw= ĔtM(!F8Ezx@yE eG^OyX/(E S:(L*xW|d8M4`~@G͏\? [,e"JL&Mb }4p䌃&25z*tϛǾz>$T'{>*?sYhFZe{^0eBLhˌyaHnm2B*Yxbx3'J ?:ÃtZbx6ѡT,uG;xTBoC"`Y!9,3 X 5ЌgEJsT Fj*S# rj?Ԛ@7:ԡNP:*7XřRǔ!iL%*f2s(SȥJ%Fd1W!&@拯ra&@wcI`&j!z-.)$0_0 Fkw7B5`Gm%S˅Tk:ƽQ5C\{L-jݻ\"C'cՔZ?+Z|!9{ZqXCH2{ZhJ { Oͅ"~y젢 waGt( iIkanlz-&J- v4KZvjI-- aR U%/I}+slK:b|G#lDA|): %_!1b!gFM\B&ޭG#j墳[|*Zw!Q/5nGA ľwr!PG-{ лuwL@L`uL-\R@^=.Q 6lڶ,>/Kz+)hT]euyIjUڱe܅.₡L![cg]8`U&9L\(M' >LwKrD}oGX$ոH)J&GۂpPEy(j!~k a3 2dԳ50߿SxN (É|&0=6谼%$G❂ ,e+~.޹@k`B=ϱ,; 7 ց2^Eyݔr*N6#AeS@2.B* "g e7V9ΉpOW@Ы*Wlv7 ^v[h+lV[⁎;ǀ\(m^[h >}_y]\6j^\'RU3fw]@Ƨ8>|1ò;47y n5sCa{RVM߮' Xt=r7׭\}.˅c*!FjR9'el ٧o\6YBכ%zM`߹QTKyb |h TNe2=ozɯS-e!rCF/sC,ྈzn0h:d/'lQPʁ {a܉iVկ‰!æ5]P㞈!YK]G iCZnbtB ȢP;ر*p;]M4S W`(8ZR3E÷`h jtJ257(@fFtUQz("TD${QD&JF)#y(e$ JCiIQz(*V>WR@&VR.PeB)!(D\'RxEV\7.T/R=JOP ?m2C)+W|y'P %՚QfqN4oWJ ZBғF)\*:RTkQb9(MM.t3KTWb{"JP 0={N %Պ~WQxJF˴(e<S59QҬmsNT%Z>&JQ䡴Z17sW J C=4j^zpTcZe޼! ŧ˲ƑXX<j$4A)08Z6R eɂ_OV <)d;pt #&or9H>dBc&ܧӎ1>7+]^fwgKԐ(;m)=M;MAf詮ؓﷷ=8CBliK|eOşIoݴcQ?l~"IjE!/:4֨| k1me.JԽk4.鿷ipp~j Y×&Os c?s!J.QRs?Ԧ${R[ojWGlOME}YݠWOBU6QOoэt˺靆}s!*gsIk**Qc61h&tc9\8/5+Ph/\gh`xmH@)#`<~]Ӻ"n6.x@uOicD ?~U~6\p'JQ##k0=3@zVHˌDhRkCC!]-˔-!BЃm϶ǩ`tze것qLޚ]o"նE|hgbAKɷFĵgN£r::hDi {?7UԼEI8ӣ7oFZ|GOWp`:.f|l9`վO{ff:_Im'b:kdUdlRTZըmN#e&䬻3};k 9ط3S;sd"=Wڀ|k =%(V%#g앩WKxW0;O59ZVDEݧx#A][vhO`Dsnk@D [6>_m_׈z@B|u_j$t<lMm4ԠWm;ǭyVj G,*R"6)> <,Ok7E$Q&a pIJ)oԫ_$.JWwfU s?{[Vޛ{?lLI&IAU)o`-: l( [뉱 LmZRB(9PGfci.]dPdr򤍆-@V(R(!b`$NV D;F oxk"FB 5r_Mx??ؑ4CٲR4uO}]U= G-iĘz>|~&#1=~3 G?]\̯˛g@@ N_I|Y4*ۻFߏz=r7$rUOq钜uWYG "auu\Wչyd+$w5fj<3@\ib tɝޚJ{Q.cO̾6O3ea%/s%D{Sߗ쬋Wktwb&cU.ZK]=kC-KJv1Hg<╔(u =^ŸTdj1L:օE#R'C)ˉ,P_ւa@NL\|{~=(!VeYۚeT6T')å{OpM9N)xT8 x?&Aɼ@m\#fG 9p<(Rm/Wp bB).LE|eA>1q`2|13o(Zcz"PabŨiH Q8xsha NXE-B.hQ:G\2"t+ =md>@)]iXGU:.7eu~/BPJrb"5&T "77NA&[(6NMSZ 4b<V x 8ħy+gv<vs~ A@Fp&80"5g zu?ܤ͕xf,qD2&qU|%ס cJS!` 54( KT#7eOC4&Y|N!@.gddۊv3:#{F3W~fwᵤta*֣U@I\R w,Gc]KK66D+g 1chĶrIm3Nh3”_{$ 3\o @;g=Crϩ7NY7%)x{52NA >qjkHmXυT+򹠇m^}pn=ť2N!XiS.܊hB0ʢb`oge8Nº\V)9b7ipN 0UMVH'KkJP 4܊ܟ_IwKnũD}aVL^Kn%k):d]iMLiQ@!y1'HJ5APhQbyig ɥ 溤ZqsCuglW2#05IXQ.64W٥QTA!! -2PpS+5Z*J>X MH1i3q*zl3UeBH; cc, LNM^ s*.Km]W]Ц} ?~`Dx~dR'՗pqM :]Oc>ͼ_wzIJ|qڐ*UN2O'kV)_׏f^}ˆDE6C"ŊN]E KMY<oe1n?d~1 ϔ$ZCyM:\Z-U!I}Gw+r6e;oۻ%4Ի57L1qy7D얨1Q[w[R57ژbګ`A]Q2xS)nQafli6tZқI^0/6q,lt0sJ7$& oۣ׽ܑ,z󇏿/ՖA0)`2PQJ4JC7jv$<LQ^=b W#3;Ok&L5p,i$E!8 UVpH ɣ6.IƎ PX'&k{]aۄ ^`iѥA#ˌ;,/sE$Bšú|Ghx*Nb{(ŷ!&:qF\),Zwh#* pOR:-z%@k?i=j.Z 6$pd`zh k}chڭMr.2T`)r2PhQ(E%y)ТɨXC9v%J#eBG)+DqlIU?S(xtr3`H#(@qJe^gPdLP(7FfոC!X*CT FQhP^T"ZXLktF g!M`OIfA7\IeEwc#bɳ2/8(9X!T"G2㤤.)/]#wH"CaZdhj?{gY)٣ mzOQ #GM8*~r,\ȒdkCdyn$e|Xs$N(gRR|O5@ OAl*\L|sGQ%[*!pn~g}J@^z__Ὤ2wYedTxRakM)D ;4J ڍ)ww 1 /I_J^&={6@ޝʯ6w?,!OLu+XHSJbIB<oj!"I<q@ݐxDZDd`J|rmP>Oa7lNR=ݹ- KE^5OepA~NàV?KBHBȔaYlV!h{ֈG+|fP_j9a| R+v0P TrruSm ^w9'b ?< r7 %&[ O|!;(Ea<`!ϻȝ_p|+ LƬT,'D)g:}N4py(Jד<.x|6ZE{̧d|=_s)XJt>\?y-Y^!vI"%vX<\2eRvgMYx\fBh$}Ruwx.kS1_-"л57LR8wKAtREH5mۻӒz&!))6e5 ;k>Uk Gj A 0x/zrM|>oJc]ۇyeG+ٻjN/:y&%[O\M|qՀUKbt&bJ %-;8ln(&O\rky\GG:Eɒר<~Qƃ*;e 4K)F/%HR\\ ,j~+mƷBAp9[(π|πohp5dpwkk4e7Pʐ(;r>{Ⱦ`Tt v[UFsy,0$L2J.,Ԙa@sD.0~\f*'XnZد7;a7ڑm9}ݝK:o9b3ol$cG(7۶[}3 kw̭sX̿>=ٯ.v񟖷(tQMOiy}n,u߳+_SN= zZׇqܟ_˿w.z߶) {MQY~%bU@ k橆4ZTQYizbiBgBd\H:(ZI5W&Up锩 NVv7{RK C)S 12Jj<{FPJPaxC)<fWjI{g ko. J Ci%rFQ 8 JUB 7R z>H9J) C)eٓ{M(J!*F7R BQ4WT8@SRJM>Igj[0JOI}+{7JA(N P " o FQY=Pv}7{R .(1J =}yioX+goA0VRKp1 P@)@J+ Q0j$JCi%\m2f@2-V3MNK}+aTq) <_I \3F)a(%Պ!=-͞~cDiX{ZF}4V%q^ǍR*ƒANK}+BQ_=\u>KOK}+uFQvϫquIOꯤ{9JN|I1H}=a|uO0| äSi2{#5q0L!vݚ2M ,"VVʷ+R{n$:s2Gp9||WY<=M}(FkZ߹?: 1 &(߆^՗ߴZ6fՓ?|,&|~bi{&#/$,+k:P5=$p$ 4I|{ĎG%!0w,P&ו2fQҰYT/sY`m& `&6Vw*6 >`QA'\F7>@Ki;cPcPSd#{;ww Oqx;y1FB>K|+|)mgwW~Z,yiW4v "nݘv[-ja72YRο,)TEE'v)(ٝjag 0 R[7.: A&:\Z~KsC~j ڬ=.,ۚgjm\u6VWUl1ۮOUg?Մ` tv/%0^KlÊnsvNU=o#䥂Hq)<Ц4䌖ȹ*r"2䨜KQ,sF"+ARS) ;n@5) {cRvG|Q"E. 3nL9ˤ$ #f2LKͥ(I阳 AlM)Oӡ,bf#yGhT0L^䯦wvKk18~Xfw_S"{[,UƇvĄ@1apJ(/$L\\QˋL!B,ϕR”?{ƍ\ፆk-[J]rvqT̈́"U|8q ) _ %I=%m$6%l[E "T<&|W@o6.F2CIJ>՘Z% Tr*–XĆD13 VA,NEHFےCg; _cBƹcܻ|QP`S9.|RDgz},JQi3a[z~yG2͗yLQkw\~DLFv6~ODFhd,M`q,Oخ|1 pn"%/{W t AuޱnTsy+1t>PW?:4T%^DN$Tjm`n]u15܈ Pmb hr=>Jw!c?~gт^*Jع~-)E{L%}1 A+Y$ rB޻<J)Zz**o>hS^,Eٓr׀R}O!]Ҡ#F%M >WJ?FnCjF?cN+ D…:q5iQ%JA3\krIF%,N1*ѯm^Fa؍~m4SCQ0ܺ7DjG#*3k!K_&1KcEĔIsLI>^iy(2A6 RG]džGlL4(aQ^qVZc`Xa9ke`k0{}uHlϳ~38&O2;~Ԏ柿vMR[* ?]u,IXʨ$hL"J }&SϮ˱&*"cGy4 gEyR|0#q6-+ܮl~f!bv/x]{ڻ{^^LA]o|{9MP|j# y9dio8^F9u!Kgr54}_Y*)/jE!v78 dwf1b4/{v1{86?M~ջ/S_q "iQ Ag^RDQɟuA).!b98H3(pc.88ZfhI*hM9 쮰 Qn(axu!sjxԅPD]C `+ADIjQ<ƑhPPSB.<#fm#n|'ytTBn;ߑS;H*ѠO^2f(PftU(0v(=Lf@6bN7֮,DOK?|*LZ wn׊G=ɫQA軜KVF<uu/%h(B¦4;7O -T݋TrXkw>Ehi#9Of(e5T_R;}3cܐ|"otZm_Kdwװz?Ǜ.:-> P[C-TjBi;i<;Tr_VBXns2ĜP{eoΥnsk<պ=]ZDd18_&w,>?Gtb bAj @ \}|* &8U~,W~{ݮ(!ë\;j'47tuuUVPa1yU1TFzuY_kūDSmZˋ꧷}xc5%&aڔ pppÓAny_9@g䦷4V`!>M7@q{q3Kd9혎C_6pQT1| RsQSY{ B oz3d>#_0ꏯ?y2p:84EqLRG@f^BiɮXryCsy%!Fi< h bΞw#L mo }uut6_"t3~d ߦ{7 -o9dJFv6~DR([㩈S]ߛ,y1Xy6Rv4jxxJRo]kPPe-ʎEx֖o2; Aˀ~SC0jb\~m?Q=jl{=>p`V|-~U>q{\5A$癁q*滻g!!E* F]qe0fi5"nd>@d3`6`F2L}\:Vo jPv8a[?{+(_%7é.nZ\;Oi:/'GҿoQ&~{NOļ&9r^?f(*kܭqV rOEVEc7,xZ:, |IlHd1J!oyy\DZureȹ0m 63'ýAgI6%y5#QXyi56#D[QQg16Qg4AIg39oL -G7إ0S\шjF8K DAWy "]w!2⺓Vjϸ@I@&q ̈c/rI9d%ݣ^ckBdkG~Db +LZ1B*J%2qZg u)u< epgx*;rD7 \,F+L/5Q@PzGm75GJIvp4D^ i?Y :dA5wǨ\jn6ZGo5Բhخl4Fk9 ٥^n Y *PuA>\wA.\wO~8h=x:gb[d<bIMtnME%Qtg$VDB"!"Fgpӫf!@-aLEfW$%5J?{8y9W*i|_Ҍ꒵YkeyF>O/[Ε*D䢹.G!ԢZ˥ Y*sڜ+y+^qv -||D)lvc㻅:AhnW n^@i/lt t6ul'&Ne5^*]P ZWQ#FFOA ,, 1 bcu (gT{y 9 HWgU3d5U9vhU^2FJE+TGQZ<vqv:80i7pgYO <.:dvU׌8*p*bcِ.FJ,ZwQg*q*AAI 9J4=~l%Fغ1O"O­ nT3RәҲ8,113e1O&@G#2f$BI\"(B 1itvk8D"E`$6R0cvʘCD0NZ0bm0"2'e{.w~BQ(cR& 6K.%c8SnLQH'3Zpq8mYi_Yhi J[3rXI6M,*} Hܶ Gd[[/bynmFkCx҈;nFΠ$qI5ÛegrIV0'%M@P{p0A͖G"c1q ޵6r#E/Mk~{Y$7!HL&Wփ[[dZ &Տb}b,~%%"&bʈNg0B˱YV0)J bqWm*h〾}Lfla}0Uj7ws#xͤvD뉗=i,qTl@ euC1\[ԓ C1$$üq":W v4f}]A @$Ûڜ|<>P$2%H7@"*!># >CJ2l[G޺*;evhL4CEY-=xwCSf*,h$ ]4yQ+賩@)0bT&"*KrF 0ZLH(*H?Jȥ4XcwP37%ì2{V,8I.}x&1:P):!o,@e3 1SU;QJǍl6$EAj&bg`O{jM8n{NɋwA 9Y0`0vmF+LUs?︝ 5Znql^K\G:s}T iUFmlXQ+Ql_Ǚ n՝X/F R$:um) %IҞ8~x轘0yu}+V]Lj4(D[tUmgw6d *.8o²(PKW`J'(Pbr~7KM%=uǻRsFUth0h_VKNbH +q ]LbfI31yQ 3BLNN?5eþw'5C*{AOy*J)tv6\Lɒ>K&Q45_cYHs4*1#ez-m\|*0{/|߱%M6E+S8HWY6n/W nS z_eB>s)оϼ3z7_;x]SkmPWB>s)is&wՁw;,0n un!,37=6%$9VW*4P#\pA㕦s?Mso6=Lcfr#i f-±J=G2:y;UUeT4U7QzEtqw9o S]&V]9bykɞ?όPl+ɑ1yp0B(9Kb0U%P<Ӏ50JUc49F*D]>IZh*<ta|(3z7_;xs"(epn!,37=rz̥lr[u~) Mxss[ebhEU5wpTC񲌷2Ra-~\a^n6=tY & '"]䂟V\cu?(&+X ~bs5Z'aXBp'DǪ`,3ǩȐJγpA q7Ԟ'TPB&Iޏ'X2$G@2!UJ3]X}jfڃG2RAs}2"D3V_ܺy72oqThbGq)H^BX W;\a5D#ht+=<&d߿RǭS: mFQg&ǹ2&-/)O|4Y,=ՍMx`g-GZ3zN|1Jz7|^w;yL3'ceIX!OM>hCLs3Fɑ&~MV}KE@UNn7a{T8p av;1炶^qq@O3֮3)& >*璷Gw%>jZ`ڢ[nGugu+m󊝨b5+21hE&t!21/U>"-uwW]9bk(ER*٧|=΁NI)_ Ͱvk]8; Țp5;z\ ݄Nj<[U1m5(Ʌ{MT'O+ds3l{uXvUs_n̷u>v!M^͑kt\73XƆ3H?;y0"_GA7'FSOF3Íg3\{=Q)ϊ/fm=-o+_B~\ή߼DJj}Ojf.l{Exo[0%&c%٧%SVUouykƑ[ӛo ᝬUKX'h9,R\%MҌN0cM2&~Ux&"ni9JT:IQ k$,)oRXa"EyLB GV߈j5wS+Rrq2x*= 3.T/Q[8v ܱł(ٙ~| 3ɲPdȮn }ha5~9Z:>۫2kmceo^?mWxx wEa@ƟbbmwF67dc/}wX wr &a/).>᪱ YwahBQaCiVDrifGȈ #iK gYIyħ+@EN[';lZX݃þDM6\$za/I}/BXVzVJRiTJJ_v_j%#7+^X)GnVZJ-^kD\\gHJ_v_jEK/Jݪ/Jd&VGRj^rWs|-{aq/V\+QlLJuJÈԷR+J`mY)f%J1 +}I=ԃ^RfTU' +JK!.vH/JsÛ1I:axG rKطb/Z!9jue\\jŝ69:0jMUOS #%'F=lOzxW6=o/rem!XXuqhOoYK<]PCWh͞ }ExV-P=&)ڶ'(+Z`/щVC!/<\p4Z$"yb9L8h5]{, u ǐ1 ^H|EuNH{vl:fMe؛>4>rC&&QIg]R:e>{Ȭe1ł3\~B7S>>P/H̫K㙛ky{^F;+=/ wJ'ExGUnM[t|9ngr]Y<|-|\4?O RN~ׄm9dEt/ǸbƙZ,t'8;;H/\X)E86rDL캇Nޗ+TwMNr TWwiA 4Cal0ԇG7hƤW X;١; ]#9OV"s/!>oZePt}_oѵ@gCJl¸ﭜC ,k_~I|?Nᖿ-8Ǡ :)GqsP$b&CҌ$%cklq6^9*:&S;*$F'kw_9t _lѠ'6%uvI_7bu9K C%̼r1E w7_9t8P6KvuY+Nj)7f۴֟7;Mo=mUawYY/KMELʊw٦rVre, GM=iϚ؞&e)Hkw|Ӭ~kiX> :0LA%e8QZiXsn;Y[rX`Ŭ%xsɎ.^G@zN o; Y_`T 0Uk4ECz 9pyv!. A0 K \!AL|9i+_X(S<4,CJ8z fIX8H HA3k1B2IHJ Ί8($KΉQڭ!ˌb' AGˊGv4)K |.h[8mzv8)ȳϾ $H1כ ؍}+[?ѨޘIo>F 7|^;߾3~۫'c"f8EH w)(K! 1>,:\BVEXec 5zU/BXJ~m 41$*kP "#B(rEBy,b&"娀pCf\fB۳qivg(Ow˹)Wy5dKt?3eOUbЏnZ 9}u|WP?'j ""D"{+F?+p rO)ss;Bk3˯WE\y\X'E=+0B0pƮ!;XP璧͞(-X/*QVk:TS `LY){d/%o&6`hܱ'j~ėrǚJJ/J%z.Vj+k[xJ_:?XkR0RH_AÇi9ffoo+I ,jyEܧ:] 5+Gӑ\2ǬՖ-8CY"r(Q>Q˘cF]oRڽbPkަ(\]>ǟ\4_..PL9H?{W8nrSO;.S|ګ-fxlǖ'~ JcSh4Ve+cs~]!UwGjd)YheާCԀEw_07G%;>Mh85t?~'x x:-*`9` H%p0 z%%Btolz BJj$ct=u9_\V;?qNiף-z=W/ 3%U UZ ib_@ "Ed & ,T4=<c[1B EI 0E6)P~P5^RYvlOD2N9Na}ʹr ^ èᅳQh̍DL>vqۍ1-l(wL]!zC/Q7i$ʺ صbzCF(t_k4 CFMTv>f\+"'Ds̿޻|-7~XCw?a_{|ه⏳_7OxY 9|kº:_Jɧ zF]tpEaofJ*ܭmUUGUHcZ1U!p7N0BzҨ Wَ+PڞT CkNM &wvv{WOqݗe>'Ohɰ;aT2 &:v1{Łdn^kc7L$RI4&u]jJ+Ól:laqj'J^kl'jN ) fNוZ9i^kA͎KąqхF8;m8eFKt6RKT #(q.qRVM[d?\_6 8.5J2"꩞s٦zf0{},s|}_ޘ^nq{M>%4J _֚~?H}͋;XsdwЩQn-ìe:zi73nUQzФ# y&bSɇYeuDA>w1 Hօqݶ)9#ٙZvz?Js`=OObssX>HOW_2<2dQgo--DXR_ץ瞷OyP[gT+1ZJAG9^A[[,QZ!Y\9\0:,I|/і{-jB[ۻ,/ -4VDܶA2|Y+}%x]뺰;gEo𓭱N'Uʀ ϟuU̾_1ѱˏGZ:hfjluLQ|Ay(ïUmڣe J*$ \X[8 p ;D>߯~cpɥ)7W&ˤΩr߯#@hU~"]F-6ޙvy[tŨ߈7 ǐî0Du.PRZNdIkbu6zg/Q!{@mE:_t}POEbψ;^ ?E?ㄉ$C-jBEzfZHyłc$/GUz[X`Lfϵ9)X˩@c,W\$nr[{C|͈QYE4-h$1*e3BfJ-8! g%8,y$J1eY|%d7=osջnWK{W ]8zCڇjbH1傣Q mh[0cT8?XdrM}9L ts^+UzG:Fc7VPB۹H BP|o#Bkc$\^OuA'aТ̄j$X O)\1#grVvi. \d)?0TDy.et<ݿДiEkg`ޱ`i8e~kк޹X(.%L90kMh~r-%iq̳y)9׵"(k1l肧Oެ>In ʘBxFem>+!V 1*Jɠ3D.cQ3ZHMq G8$#X uZ`x=WBQ%~2ZNdOfI{<>)^J^6\D^h˞W|F|A$,O6T>@G[R4?~1:gnQfy'ka%vZ 0|$`Y%nM6tԛB@]Q6{GBCEzik>TtCmWȷ2#G_h%#EF^^T o R+8֎kғvZiMT(FdI Z$moF+=dB޸&ٔ~ޑW5FnĘN3xi[z+ZwB޸&ԱfMIbc:hF"PO}{!HօqTd/.S4D _e.>BJ_^Цs٩WGa<=6f;# Bsk`xiuߨ M./)S 2rݛlXڛ*m_QWD)z gwůW 3Iu&ry_M.0`ko7nVk/B5U]P|ct O[|Kp~6 8]P`=P=>}ekkuYtA-ErV#]®JU~ PJeRc(%ej j\t}. r\%AWF%\~1~uq Mr1-s9ڹ(ghBs<1=ؐaS`pRMI+zqtg`'|o ڐz {N3 tH^^G JԉoC =T+{<DŽ6Ѡ`A.# Di" W-d)Ʀ ω((@9*Es╦yQ@ae,LeZk4iQcۣ'Q=I $dډBbHfe"Ng2Y9,ݺ7n5l GlK81n5ǘ;e}-)"Y`/]Iz]?wstT$A d"KأʶX*y騥J$C%/aJ4[oR)LGxOY&3rz09~a'- Oi\1Tz7cAnnFǘNwfIԤڢLYqr X>7Qixi+Je4 9uJVdjlo1Mv#7z2\ko0Sgs4Ն!L.VʨQ m),+*OJy5[4#6c:3"gR+e92|Ta>W@ڃ-0M  l]5fAFT(OPqyz[o-B\攁5,˕cȂ!L f2ʬuVU;B$P )h`̺H{՜'RZ1Mmae` #@.Tac7VTϿ=Q_ֹ ERZ(Ue\NQ8+h!w =222BjG)}`nT3{1%BgS㙽07l¦`0S!feB2{!(`Fb_Q-e jT2Bpel쵱}7 PEwd_@O|YTUv}aK]z ͳ6b+(aI\l_JX+ίqHd9-rW[auI氦RPD[ L[ð_dnjӊy8&otT m=Gn#E/\݌FwP!RTRq]݋].$lyݍ${ϗ`F)p@r8+6h4{eۊɱ!B 3|4#jOq88 A{BY H볐|n^T UҪD{!mjga@80%JP9Fqzѫ7RʳO>h0Nbkn5 %ԞuSia ޵GJm=֬I5nXìQ4 ylGUbn5r[!OUF`lDמ43󍵦-M=tVR( ZLa@j-x xH*ވ>enZ]6 dF>)Z)8wJ^{!Yb؅~YDxzGs{`+2'Z3Hճq^q'U4V) Zџe{2 󱷁EznELm&|&; f&`& h֧_>֪m?UY>nͶD'H>e~E rވ=nA[Iwv s:{c*]gqIMhk֟586h0z [ژ帩55 NjI^pAѕ 9yVmrzEwIghу wrNstLu\:0:Fg#(rBX' R2S5(,MEbX'in8IoVH瞤Ce0Hʡm^ѠRJjT$FğNy}V|`j+;(|=amF(h^"}(Z*b"Cio#IbE[To h)kv`T0%`ˀQ:P5XBlסf樓X!6>]`Dyﵴ[!28#fˆK B+7pPzebHK`K_FoLb'PN{[WRJT,) \tpx&Dmйv~bW`/Jףh֦*5b4(SbrqʚɏԻ*Bt+ :Η}%n[2[ y-bNt_r ѭꔾKt;,Bn2[ yT^\!5,Wz>& {!1S!UKsVV2+Z^jy}qT3W>{#gpw4ܻ3M%V z ftpd!_ X#vM6T^`UUx-q}jWx}G>zuлy'%-&_"7Hޜ2|_JEY]=47I޷oV'[>7~|oW="Ͽ}o,>.y?#gO9Β-n 9 j?Mrz&[ r1U.!ӛ&pZ+g?wRu0o *7!gj|wu:MKBYyrMSGjtaՖ7_iM>iT|,j[yJs^e*>ӢuS-?NJs xX-Hu 1M 16FUY+0E(9o+qy0ArK%1df#䒥`k{ٞ4MCtl8PT#DH6AUS:XWI[qLsh{JWmO).,@ V쎦w5[ejŕrMUm4BʣҠĄmcХMؘl![XFs5(#͋=FT_û%%:v;E|Q$qǟM8 ZO=eTOY/x1ϭNt==狉ݓEL4zJezT- ѓn|jtmO3+,"̮HEvǖ3^4 Rr$FX l?nͶJڅog)+Z X0gZyA #B7_3nW_V,vV0q9O;X~F`QPFPf#* k-zVEqBiXq$Nbt*J'UE_H8Q4$]^! UڀtȓZsd<Ǭ5BXlY-3y'18)$yy$?=8H'cVl[f^4K1sWGIeKWVke2k)jKY}e5 4|%\JKVX͜^6KADwM Xֆ<ܚ0 Iq?Ym giةة+_J~wu>Y1)KpѭOWwi#z9IjϩfJ&RYY9|Xe>>6_}CfALb7VUyf"=XX3Ex60Y*!VW֓`_s_j}Kp f~9+h7Z7sN0+9.۬9ga0l1ap$ wF+itÿVM8/Q56hS_ o "uWI}*J`B*l-ݿ^[ MD(}SW}Ӻ {_m^-~N632G&-/Dt?ZOջΠ6 6/Mr=HYFY TpKu2^l2]r͞ ֶԾ= 9:, 60Xf3Rq&R7薄Dju`0(S}}8SS~Z o*3pwW-ո.OxǗ:ûDZXヽ݂`Hh5Ġ?Gr:qjc0i^Qޱ0lEFX&f蹄ct5-- sg.1oj)D֞u)~w4PbGX"7韴Ht-V]P׬?`]ՠCԦwYcUz5qi,SZyZ u2N`ԠAmD) P:BvW>VGY5G6v@atpI,Z fЗN_,$uB4V-B%U3@nOu] feBHH\Eb(q,ptt mmr> F1<w%UMBC#^*mMuJkH@`<ڛC_D%!]pmpyȤ3Ή H!!] HLqK432gӘ!.%ʦauM:@ V *]moF+>]X@nFY f2r0:#D|zInnIɦ(JfS$e+ l2c~*01KD|bLXʹMV1,Y`u!h5sDh'aZGZf'C.< lȽ $66OuW[.y}- @c2~PdK_f/xH67(`Bx{ >fVy\m~=Nf_*~'HI$Y `ic鏼 R 4J|l2zR B+ާa-k~JVxô>n=>ŠK_Ʀ UH#r- :Ų O ̺ww[Ap dO~Z:# ܃Z=\B yඁBG%YL nJ7?>$Q'cAQ=h3Do?=LfAO߃pYDj5WtZFQiVTbBz6J.ϵŲHLy6/'yl0Y ozb'8V"ϓ7Lj'hE l-'/;vA@9?Ϧu!,,waeSbo>ϛl%7|BJ5T{g]ח iНƿsqˆRÄ) ƚ=yĜALYt1l &bq;\UedrĘFIZN?i bk#*s.A[`x/"yrڛ/?{cMVhcw~_WkLZ]]M*z>qL@E|n=@&b @ŌjH$申+b!t(+$H囟(Tͮ.eZ]tX`78=%5`208iه=Aj),5!:'mRe;@šuU3a0ፔj-TWfcãfڼUj^V[}00YS*=| j// 8h;7Ŵzhma$j//Ѷ΋Οn]b4QV[#8wFnORYh2G,.ZO LNv1y(/ bٙvm۩3 ;iC$DR0eҜ%4)CM$5 E,CyHY[^Rcr;5N{i/5NQ%4,Jxq4B:HYQ"e8M䁩JHB`CNyRx@LX0iMxOjd/yb`UN:s S]kqm))z/Vкb`CoX.iݩ@m[ yE,кb`CoXW-#X|zq kW 'w=a=\憗N8%T>.ұQAzn/"ԍЪbȘj@+B* J#&hD1&H(qb+(:X.09Q̃*M5_4!Wg^DX9iHR R4HDƄ @DǑfjFb #)/F,M7T'$#$g( ”ab("!_zH677:nqU1|bh:Q;Ƌjl~:S#,gJi5L꧍:>6I[tDb8o]f`ӈQ+*'$ić}w/[R'Jrf`fQAf49vs}Q4n /)N,'+Z3X VlȩۮJeH-d94>]x]YM˘È#zOkۏ "pW=۫sjbqt7WMAt[j_-Srz8)FP+r1tp7.@Ju$4ӭj1NQl;TC{=-$e7]U2se)Tk1Kݨdh ntuP+pPOæ0dT[%-OXhشIj3RLzcE^:MKA^V0T18.KͰE'_ l )~åwVU2BTk#ZM0SԋSS[7VZ7_ l a%Yy۩*uk!^*kĹoTRw+x~ɺu`Jn 0S)n)ӧERb|O*pB^9dB ע.yN92D-,Ω18A1795&:.(pwl5_,Ewl΃%acG}fs5>Lo<DKSI-3l>m5 #Y <|DQ:^`#f{&"Jd]pYR3kJ vm R>GMK q.½`)Av,ͤV\Šf]܈ K)ci&5T]Xz,eʎLUHǤVPcAV,s㋺Heqe/,=kbnR̳T^Rk|aYh;NmLjM%.=o2K2I yV]YzLj YʥK4,ղ,=&$@>{K l_mǥV_f)av,%̰^ fLj%K/,=7% K`)v,O\f)Gv,Y#e,Ȏrt,75qeΜXر kWjǤ+\ՄÇqs.mo. 'XI&}ɑlFE%[ MyߑNwOmDY)))QX)~VZR-)9R2]K2 +}-'+=j+gPVhJN5(vJ):G>1~.@܃K6(y cUaLodɖӻmarvMy:uvw_}0zmgҜ[?xYc|q󖏘R= 5^맧+|Or7%"@,l:JeTRIQ%Yء,y*>Q0cT F T0Ki jE,\ PJNfS$AJі弤y]2C"S)fe)aƲjpZB/LAʋ,vV:-,fQϡka.Zuۢ UwY=Ta0!Faq1~ v#_VIl[#MQ\D>@7BSPd ]x6$dnvG- JW~3JA[Ks_ׯw77'W7pߓU/ֳ{2Bsj#S#͵ajpw54(5*+;hy7[S۪e:۶Wܬn\\W;`A//~F(Zm5;&ђtS()`,/5m$L pCsŽөϢ$QIW+'Ż/8%0yyu4\F?O& H@PAi+%JulZ2OL)$>=X;.!o|:ՒT?QVvQs\fҚa QHΕf iJ$I3BɊF5<'%(s8sUR-;6ܢJi J)Ғj;OVzVUlWZ)QXW%Dq[)xSJAY)JJ51+uB[t I5IZ٧o +d!cT.OF2.I** e)28vxy* FµNcy7s&?>/9^iySVOQg\i{x{pv$L\yҜ%ܶ=kJ8/ j*$SfyʦyDc=c<$b%gږS=wfؖCj2w˯j_$C)'eT=~(===0y5ІQSyi -~$=WvSxX9`),.򣟮][+[§LQҟ3ӷg֣˵B>fv;n!~yrOsHͤzȳL/^;ݹ w9FDZ?ؾ_sza6?E(89_z]t-ݝ $Ǵ/g.!JF6f]SI۳ѝg7Fϖ_yX Pܛ% '&PCڷĿ>”5_!"QMfJ®،v}m`W7=l3\5^(7jz{MC[X ^)&ψDJ)I~'Ek  s>F1KIhj{Q,Z,«>CxE0:sH&"lmA|o, myy+r!Ffg>DY?Z;ն;]v+vv?qYd~ n*n*Vo{7y|^ v'0sy~knht*arz77b+?,]@x,+\ݝd!Dclbwcr2wĎ݆"Fy' ޭ y&Z)iSeZFd;u~٧3?1KDwL@Wr߿ԏ=N)E5JZ)",gv0J|R !TҜ*)Z%OGQJrH.u1V9t}1i΀CU-|I,$v}ƍrNR0)) TS"F PDG-5O9Osc~/g7>*3q+ B@G &'ĸЊ#'{Uk>02D%uٽ xj"I£ '"R#:1 6h%cGܰ @y-S>>{ ӖtQjM7BӪuZMFKMuLL(Av A<Evmw?p Ȁ0 ߝ46LC]Dc8]Y ~M8z &@U3Ԁh7SITEFP-AQdxk+0=D4ܱν-ir@ĎX!Ď1?:eXaDhB^FThXm&Q ^"#leu*%s#N#jB^FTg;- m-V>;Fw2Dӻa!Dk6#N&%$_ S%Y6ꛉmJ@=Zv#)ET&PJ8 ;T"*y+P%=ERm F\ =iUySV:0cIK?MHKO6|SL!kg=cRC)\J-YΙ !,tAҔ`!hgi*L)( PI!$o)KQM@8C&y 򚑂 ˂IJBX5QFrYT0gLl͌hiHbeF#EP`\Q)r.S!X!UNuf$B2S)pcJ:0wK$j6F36rfcf ھi?و2 + `YЌY +98G8qzF%/( #FҾ#=# nTF\Aܸ(W nn Gg&L 3/<ىU?Jd̄/[I;tDɎ8k@VN%HbtcCN= ^v. MOArBxo`p#D7.8PxY6ϨA<%B'ES`TXڼ*JzJ& aUymo{p. //F䒠1X5=hmX#I ? ]x5 *T[Hqla=rPN-ldMF+ha!W0!0+da ՟)w?vyw}0;3nl*C kv&E0Th|Qw/ࢢPV/8pw `fP8U GFPؾVO/%jќjd`P/|al:2.#^?~S#N^ y&cSdn]0$4b p"FOۖ1: y&aS!9wx{@+Ӊ#Ļ EbywOֆrx$^Жa'xDNsj}xKS{uu4s(h){sKewP\}"}2ݭ&7R{NF%-J'_f\*s![p]7l_r; KUz?]ms8+*)\\ںJ*3/)H(˲g6@6EI P$e+S-l<@7ltߌV8hK'-W_6dԫνl syf$ΊWX/:v,2sDj=kowt=js '<8> O?!P Y'> dFjNn90{F x y!V}Vfp#p"OrV z@޺@VRDЙ)d|q&g75T&R뭒&#] L-Wd[f>[Ź$Irp$c do](FSJ:s!Э"؃z+AHH7@)j^keJ <4#%6RseJ"3T6 G%D+DJieԒҒlm^v[cPiڭنk (doL 01 72~E oycm0Ο<;]@_I4<9 YWy*:]ROR+N;QRX I=mJT(R`a,jykRORkOf)0"l,EJj-ɖ7KyGR+j}b:RCԧ5AKL! XӦ(ĉR0BU.'KXZI@R X"у`)z|\䗪YIN!zڔZ)~R*XJe礞6vGRX[}/jV~5V_v:mʠ "%ugQoV 'SjEgHj1/7yf1znd\}j6uR)J@TJU"HoJCJDmC9]넿^׊3j diEU nu'A8wGK>|S\y6_S\M;*"؎W{{Y~SOwbO@D|lcI>!1-bA՜nO ZOT= :&Uw ڣ@uj6C9j]-iTJu!# f)h3_Y*eOFe`V^=p mWjїv5pTI*v6-Uߚ&&nifsÔ;<(2'NY];Z8Y;pA. &D^~&'L)$!jV-.42hsZZyIC%S0&PRPto-9!*ܧvdتՆ Av R TZ{F6Rb1n f?̰m J85PT( VoKx(ʜJ`ΕoaQ@̌5V PPEd-0h̹1(NWQtO\ίo/y?.^5_ͬ|~__j)DżŻ[[}7L歛01&l\_?Ͽ;sB>fmiʜ~(}{W '\QΜm-3(_ktsԳOWW^~`W)kM F1y6_?t IA(HűiN6bX-Xw4يq3pϝqZ{LOK$Wkfw_w<I)U!,& p daK{v=anA :GW9lQ ߂I7[اU̮UW58&yjoޠkuMiC2=!D jj16v] J 8,a )(hIU9:Is{# Qy)uY.Yhn@*pJk>:Kߛ;`.?kBz"F*#̘"F)&K"{G$Kg䄔 F9DjꯛI%nozqFY 4} Ƙf"Z+E&wQS%!*hQhH321Թ?q5T7WLoSK7WL L).Xx VmJ;c:bb/辤`~z**Ѱ 5<{*:"t{S"*&>rwMjMԓԫ8Bk=\ W18(!i-^ wU~̈>6wfKJOvv[{<磧M'•39!>/FtX:91k>Oi[Mbȁ>|(|v^-ZS:r\k):1ħ ھhٮ>*9s2k] n-+g+ea;-J:-?僋-\hH^FA~w1âAдJsÁSf(vb$A;# E=ۢq=rv#@_ap} 1C^9E8%}uSaR1N3b[.S][s䶱+SzIN!:UIũ-myu+IjfD*j)O![yTN=Fϼ%.[B:m]BK[y0L])Κa/? ;2lB.HC.+|K P=8&*,HdjvAR ]<=&)m!21&l43%kT70-ltV/F+la, "F6"b'\P'aj8.U.z h6]0K%PNZ5`J”C*$|Q~ 1M]ԻICM^~yw=<7{Uv4O̵ >B9OE67rs*+"*Y"7oέ/כR0Yٵ r$4$F[0i2\btǜ2HfJff̥:X ͙ HĊiJiiZOɺHtNS2v]*)N|]%a/E\R@C}[R%P!GNizoN¬nw־LII hJgI@J:Flʰ&OPV(vJ{nj[l2xqy鷽nߣlY׭8F%N{IȮ-BUY)v5%^ꪀi v 4ҙaШ>K5RVqe yqr1IZfs0~&d԰)LCāR*-EKRT\ +)Z#J2ᢚr{:K_DqJ["!ټX*l*vYAڬ򯙌YIǐ1EÏA8kƢّDi>OF+{lڿ!gzR@Oܙ9v$ӹ>JT!_x}'&ҪG5(d=N,fPdxb`&%0y%#e𖋅'/ueڀR=UfOm⭗_O~xu;vS\t\G=Z6ѠMw`Dt8\@y]W)u_-Qi3q7/GdVtX./P?;4Mz=4{h/Cs{wq37d rfy4yfn%w\bhPX?Hg09|>[z.iz+C-k;'-gWrȢ?^A DZS8j|UKd*IPIDQ熃N3SALiJR4&$`|]h;jdwpBWQBtb-|6?G鐳᳻عF*'z=Zgj;.Fٟ,W3' Z cۿ `(k .S7c?0 Yr?Kb6Tg4Id+ϬmhUg=þ 4p |0lJnYW@2-u`ƆR`C3I!Η1p:ҊLAIfS&I*A P$eoP4p3NfZ_Z{i0M/mXmcPq+cx{MnE{'ŇW飚oҢæn,5[o<^0&Q>x%,z3t%~7KD^\tܫwg@@ ]&~yj~LJ@$󷞏=zg%9_|)Cfj6;>PRƟO.TJV__Ϸ0*"@<{ztCm;ۜ }RK,40e~GqT.8^v>R(LhA<7)EHCp:(lP0A9 slt{dbq]R\q=.'͇xέm,~q!^Nbokn.Wi/Um8:Pg\ H1jrA'89,Lo<@9Rߓ}O Gjd-N@J ԠJit:]0B^nɟճޑe&U׮5&OdTԇH%/pkDC5_^ZA)qhci2 y3g J)#d y(%N R8fԤJ5 . fV" VT*|Fƨ<%tj@릳ArJBM=E~{b HYBdMId)a<2K@&LrΙ38_(*ܩj16ɔ2'::M LkpN2\u•J ٻFndW,UbHag_gq,$Kdgxl,g7X_RV닭nfSb,ZAQČh;٣5`tNs+J{*#YE)rg,-VHr!l"xymEBֆeDM<4oٱΜ|=C=q <^dqZ\u |ٯ~i_ &Ԑa53 ~*xtof>'P2n^?=d A1gVW[8j᥶2aX+`=Șh~wOwwرQ$kCipGo|Fw]+MP)O#NR8%L\b\r.>lֱZxκn3d F}.o4v҅gF~;eXgVTWM/5L I~'4XOY:r\b53Y"QWTcOHݫ֞yq4kϼ>}:3ȴ{m`@0xuVS溆NONr`vSi?~vUudo~2+u,M渦=?﮾:|{n =`)Mb"m~VT9?;B޸)N~^n)xNw4nNjNAgH&z.,䍛h+ڗϰLj݀UBbb:mĻqtToޭDօqTb.ƪF^Vϳ^U4/n0<|}Qä;ovv[]QHڙxRaĒ&2${Nu}O=Kz]@.ɸ:I\k|$hIs֨URNJiӉ5jQx2jYX)˝GPJ]0ִuv]Xh-dHޑo1ՄU 'Meym"GJW!fސeVƔSi":^;`!_S<g(P T̔CLDި&>!"5Iʞ/SGVtȑ ަWZ>4r7捄╄P]P6RASYhLbc,cMؠJ <'f49"ix4bIcqa4 Jo_ ÷y-&ZMțE.(&%Ljr[ƒYg#%YrΩ&GEN؛g#-_!o`q*Exj׏Ʈ6eD).u]jS=߹йK#+5%\IkviIW[Ҁ@e.)J)m.x!r@E fqř(e19w>u5B@(B9S(XP%b2cXg_ֺ}Y#ШseېNPgCr煑B9S3a"Ha\0c^aE\Z@xi`KE$+ДĵRgФ {A\2h 46& &DƠڣLj`i-I( I#Oc\hQYX G&LĔl@htFIZ֢YgF 4(-zdW!ybE:ΙT׌o/)ҳF~:7?1[o` |Oi<ӣb8pK(88*w|7 ^ڳ& UdqNwυ)3A؉M@%ԏ 9=u\h"]ȱ(3vyz7(ֽC$3fȕm4q0Y pqЪ* Np6M-Hഡ@=7!kcx,\,{7btHKĜPKxf\p9gUgR 6D&"3S6/:7m66>k䂝;]'i4P[͸Twjt$N(_jdGFm v'UCd Nޅ\H9j}GjM Q?jqO"5VMM!O~^n(*I$FܖhD|+TGk>c5 y&Ʀ~YnRDi&툫)4-Jn]X7ѽm QiqUx7D/IVR5 ݴm)-l ߅ċ5㾼mR_ץ롗!zC3Q̬rV3P+=g+#碕9pg5_crOJJZe)c:CAKIZJU & %t0KgR$'Q袴rI5>AYY{7,#;n[ mw6vFHp5ʫEOWӇѻw\P_>de۸|vc{;'Fϗ'_)'|zS3o0woPի멷X8o_#ZۭTMU4bIE O4 k?C.f·x= \Uh/TuEE)|ԢKB2Z{Č7 hܗ 8M3֑k'"_z7X>.9umP],4ۙx[ҥr޺@BQ'6;m{ !J|AeKƈjs(4,Mj2RUn?]At1Q ;lN-go&@e-HkHe-E*VJHcd.D:QgS9N:@jljҕ\4O!MFVS]RpaG_Ls/JUa!V@!Rv`6UIi^债0}`b;푯i9Fdƌ O G_O/&0f\FNRroIsr=%X!mʙ`V4)WZ)s;.V!-r>g r ,wבb .ua:+ N@l.YR 9g\۱9!:7 KT/^wZ}UmD e+m*~ˇFϬ-;3)JP1UR3&JlP~Es4* f821AX_X)+1V#aIVt6}a~T`Zdy>PzY<9J']5I3Iq37%F+kt}WP=faQb1IEkjZvt^o$l{*0rحn..,UP9I\O3.c!{Q a\@c=%j]&:dfHO<&1}.anwy7Sv} )#` _0nR+nBP)/(+YanjrV-UX0%L[cJ˵ yXYiRZ+%T i])0DFYB ҒZŒ$|{ay/'>OEJ~Ic{QKt;>n1vi$Clt\mT)ϯ@X䢲ìA~`. R.OJ^ 3Jh-9 7%c%Jy{o֭\-or6/#XE1%tO[T7lEg;π{}J}c1XaT!l9enȘS\'fRJ̸F}.A-Y`*&D"ȵRA^Z2kE^9J- ֊yTO:/䋲 T\ sA497L `.{!&Nk|w4*u\3"ň \[0CI1`K &R^U9mFsyw]BH k-S 01EE&5py&vjv3iHCb&#VS+~܃{l/ҳx_ӱywOˇN++Dž\B!tl՞MƸM!e$nh7`%qd/['{]g`x1#s )y7&\)n"گ{=١Kٻ&$W|PD?xW  -a-U WpEZ}YYWW0h9x^ɶ/$4q5 ,a4}G.Zb?zPS2" METĐ"1lzmQ{rGlJTCw)drM/28z#h,sccR"x 6^87Z΅Mno6S!ʼn ,92xDPLXQA9\iVew.-)#TS,Gn/4[-cI ×RkG]VFzl:6㸰MDݭm5 bX"E_EQFF줉qdrkۛ(*c95;1Dh1CASLF/zm_}>5j^~JKzFF墰&6]dfp=-C fƳ[Af+;gq}D%,L uc5N.ubSFþB;  Zk"3!~a+9cXHeoY}VRv GKM!Gs1jaw[ aꏾRfM~78[ޭ=@0Bk0!5C^y a`ǔJO|_"qV]xMƕr. /g*6-Gwů͊zpdVw=z$^/>4>}aշ=X0:)^q1 8f*݌ǀU /le69Y^nOn|ߧwL{ҪeV7/~Un'}dy Rr NB-YiNk&QBe~df:j\\uriUv5*˻@򍛨M :eʦwP/6 SsZ+ܶwOJn!,7Q/d &Cnؘ7\[[(iͿhw a!߸6lYP Epv5;vo=Rg"݉qKMm q:];w>8'2CjR0.*\BLjqs Qev3UbpwQeNla@twWq 0@$.qoZW>ݹ)o#Thk yub[uA|6]4 t8!D_;1 @?caKe9c< -a@ݛzIl(]yLyM#Ż8Rs i) eKMh)P!\b`s ˹)95\ M˦̷d-G拁~ŻaZMiKBq)nA0拁~Ż wvoBcBX7n 6%MVy|'}N6oؕ䠼PҾ*#yb4kyѰqDQVc-@&VPd6Ml $wlz-z!'e퀭/L9B{7 9DXА}x1pRCځyۇ4U/IjJEDf*$>djɧ!1^%E; *)EyX,6SU|öQ<%z::,=9yk_YwX¸i:$66;ihms E}< ۝n<9}tM!wU${rBREOsCG{mC)&saP#qHsk(|E'n0 NgԮUL|UL?&]F{gVt;< y_]-5#{7;^j9߬DAyM؂ӌrE$&R $SX iN$MR'4XgrҌpQE CM-E$"ph&j_hvva SdL?~(LV$~J(Bd7zàJ,־6s>[$OnpS[=u-M[=nU܋яS=XnlzjtYu_5K-Ӌ  OGϖMa6gez?ZŪ/]O0ˣ.zY*?aa)[Ej\xҭu(”DBSa, ׆bx.68o}s%j0[qMs~MgJSs_kOf3-s5I J-jWM u9(㨫j"NןꜾ{p) )Td*| RUp$E# 2X뉞7  <\Mo_ӱȢqD⭝r}Wj>: J#\yhA $lTZ*AMB92pB9#B YXHH"J*ab-ٟPn̈́R)i|P^@qN(I(wX9rhyV@v*ֳ ϳ S4<%ڣKjߧ# &Y.D#Ŕ&B+ŭtmYl}?2.xۇ3V-wXf[ᵴŲG^ouԞ;>I¤ 1v3Ee@ ePce,͠2Z{*qdF4pZ4Èգq#['"Qyd 9t^{pڥ3;όO4 ׳]B8HG*i~/uT;<۳rԱ'sH8~vDly ;(8e dpBjqt' t-ń&y=ϻ/^1mlͦ' 9<4HIT@T2 $jsb0% E/o;SE V7**,5,^,Y noFmi E=|ۣ˿ J G_Lk>R 4DIӿbKZfJw>gG+'}c;nj h=SyyadMt[[s/Q}\tV!dАwy3DlqdxW QuW YLUnETUFzxWe뢸%^xm]|s:`~0?VƣE%BTEPH{xlq PZ%K gLUnX(Ի2^Upnv5@)w)Ж}ν9{+XbO5tVkyk J;4"tPڹBjkG,g{[chJI}ՔZ""VzVJ,'R.e0|)\RKVzVJ-g|NFRڔ)<[ [)zǨO1)DlmY)+a tWM;+Ey#PO+JK9"g+=i+%J 3|C2WMJOJeyR`ڐZV,M8 (W-ey?3U*(z~YrHs(듭] ׳],W]8 ܞjxb}WAHuX[ݵK=(/rţ8"%Q RA=/k16޾9߅\-IDI̘.j{)}r}}<A;1T%Yx(U"Ϝـe߰rb@QͲrbIW9 +10tȂ`X▫8= #s!W(W{ސۋ+(Tld!z-(DB9W0J0U4,I& ' #)S zr=[6M#SDq= Ϙ)X0 &87WDؼ+MyrAXh0%PԂ#}*XA~J=2Y<7߃=H( y8 L=p򄧀'(aJ0R.X"XJD6 Y&B [݆Qp+a,bƧ$9B)*M3 b8/D)pL EdR>\æ*P~T̾_q{1 ,G 0,6'Zheh}dDSjD"oy~ġ_e[^ʵ!L=XC˥-4 "--;mWVghy3(Lft[xDzCC0n* %w  Išƣ~jg+ -޵rBeMH(/8yX\}В ߷zHHscG:lTW}]}4mJrul"e\-~^pQ><)ﱏ;.iUi4DUn1Aʘ~j(Uz}m\tvf 3j6kgCT95wx8Y:_xU~0_rP=Cq,mP hE.e ,ʺ*j$\[=zX);пg/ϣg{i8zAhq;ŷWZ8ݖn}y]NgFF*nR152##qC CB[RJQ%W5Jch\301JBjgR1NԶ+!}[#5QʄJ(r_JpCi!`(=krR^A0-ȕSQ*JY3 ¸t-5Q*ktvEjҷ?ZW5Jms PJJ aL3GrDTA{'۾.2N5ؔ56jY:oV9lU)3@Z1}y'͢J}HQ (pl\flCL9N肑4L}Yɤ~얱ݢ߁rp9I@&>DqH"dtwTX5CRM'p~1J = 88oC10\0a\DŽUCʺy-i7.J`Px.PN[ְ hk^7(,L{Ő"?.h:aύL(LRb EiF$3LJY!O3e4JrRUWeC{uqR{e]U{# ihyNXR"y"K&0+Wkd,DK3(a?>e"C9N?_+oj:EׯYW_?Yn7(6J|S<CePBYSS 8qL9GYDE@<°ᩙt5ܹܦ S㎸5;řmݺs+7/J}uARztM_.)8+vAq []$f[bkh^<_0&`K߻tNq`xMRHܳݭ*[H19PjaAno^WT*uCۑ \8D0U7HÕwf[\m*ʷ[pa w GMRK-T. }wo6:ndޭ \8Dc) "(֎!N k5 a1Sx\pGLOSm '@}Ey ۬doߊyӭF8֌FNGξd6rsmoϫ2N5}N@n0zyb }aR.eiW?).@L(U@11z'N|vٷw+པd8:v>\ xG{*" WxƉ1KęR'KU)aE!(Su؋Փ[ nŒd̴ےd,Gėλ˗y QFj+G@Bp$Fn6|MR&I$pl1Q67Q(tuK߮Ke:([=F[boid;>Dy Wc 2SNeBW$w++#) xuS{&G<葤dw|s B)AkL؂%Bp\ⴇ6R cr8#&iY\2W~ MMˌ@WeŅ(()18wB$DZXi4Yig2Cj]އY 4P>r2Ω"Y"lpwN&xG0~ 6\^.!.хYZ+?Vyu~ʈÇOxUM{K>6Ԕ w肻MLo}y9EwbTϣo^3q\BDe>dT "AbPePYcu;PzCP< Y̞Y;Ȍ(jX*$ Z(%hlҌ&StVi/6%3l- vǦsv_0K_^(e8r& \$H1GM O@ǿ}u~e\m,KQw@ k}{a†N܌rWU?5ϩqzBJoFO]zy|G~~B(~yPƸڞ}ЭvO> /M}) |գ48 7`ou;d''W8L<SЯADO˜4s㍾zO=s쇇s6IRUg?ѫjk zpCmpD %Mo?l_{ m or-qߞv7^L%!Φ.=^ҨRrOzsr4pt }32]{VskA{8T4 fή{n;8h^ҝ -K;G7+`b/Y{)OW9 =٫rxN!:,۷& s^_g*L GҜ Gޣ=3-c?-DZ$QF@k*)vxj6TQJ] Luj\E?{u#ijmk5Y*-wuV<-[}Ji) r/ ;+Bg?ÌRn՝|bLB:YmN( VB֔c42﬊S#6Ê+3 j8^S֤5}2ك;qBfA֤qhPYj""TkoT0ԗTaVS-@'Nwkwh#$"Egk Gb&r MRzn:pd|wqe,SO“6[pa<wӥ>n:pvq_'ݭ݂S]n !ZS+\x+(gu{2,8 @0wKJu=QZiYʺϯYdivjpdJ=C绛/v>o7C:vWT0(JW Q-lx!Ya5f^n;ذS9ezKps77N. AuWQp֝[MEyXT*j`/0 ;+n?}vVWqM^JPm]!tҚrђV5|]秂:gU::~:,׊>_6]S%>ۏ`ӇH=*XM |K^mYpRKBn) -+nEh]RZKޒHjiqkwܤvhc$N7k Ǜ&r T(gCZ+[\jFƻ /ݚ@ȅC4S~yT&@1:p |Wo6 K: ܎ߊ ry© J_/M8%w~,UȖQ_3Zޔ_iq| p1BiHdd:v%56%**o#꧛ϫ2Щo5h-6Y4~63r/K qvdKQuB 8qxReof\\E,ӱ!Ɂ@œO~xhٷw/Kuew*^,a/ހT2+˼W3{Wk7P@xP&A&`,?WSvzA I lNfiH e!BIH@h8Ui8Oi&*w贛{菸O=!6޵q$З,~r6,nls ̣'בT!o#ypfh@Kԫ~%;U%6646X.@RF'כg;[nǼisVqO[Iya`eNa|>,Gf<%dΊ?]Ftv X=Ag [u?k,bp҃7!2$Qf->"]>g/s3._/NoOGP{{{{S*tM>J`{-#wמ o4ZD.M]9]ξ8$O]gTޚzdE&~/T6V+ͽ"hwz V$jA $T#FdD9ADRRVLG&ui`V4Y Fj*QXTOKTKI4aV bB t2 RA>tRhhy+%asz VzxV*NX[*1ij:ZA[)7aVMoa܄YiN6JMη!^T5hWciZ%Z{|a|{WI ,QkQ+.?ݠGL'}sGOG DhCckf_%K_7!߾Fdd( n.W[,A`^~R4eq<41ZgID%oy=6)p.C\vk[%)) \6@RdwJJ༇ /mV׽A%`m/mK3͠V{1i@lߕ9[AzxjhA3p#±r;?Ƅ'@$ܐ bzސ@r8Bl*wRks UL8U$%RȢX[EYژ2fAݪ" =WX7`UUEm"#'Qf K'n V'.hPSa)Qy&Rt m\Ol6N@y۷/ŷtĉI"|jjr&lgZjdDuG#91s:'Q[LJɇrΣx$dF(VHLԪ8rw&dq H`\GK@)ߞTKJ?=h݊T@b-JE),i"\I 3tD*\IVh|UV!@)jX/#eɗsf>cM.VWһ>\ݗr˛;s ??9a jf~1qq/ sF|lX)]';>5Q>~{tuO`<]F+/~4QBJg)Sa@p2Ykw/ئyE:KpKN=Nq%wꑌބt%ލ7uH^ E3'idЙՅaTޙWbAgUU`ԸXvˈV-K[~>cymdY%u00E0:%3\U$8(o$]MDJ r]D<θjdPPΈ.+1K$'옦Ns8  i}&W\tj5e8n&~WLu7s-n&5*Pc5ݝ1%u8҄qacwTasu!(N3̽ ǰơ&:>Dkrd!Z5a}SQ.*B=7556T+TףOSA+r]b~}\dD&STZ~MKT1aMe23 ϧz`F90y80 y&ȦYGT!U[שμ[@ֆpmdSZ >|ݜL[SNgTn="agޭyXwkB^ֶ)hyw[]þCqط8%< n2Z wWuJ gC6P8;Z!Z)`B^ P }Z)`ª8)ⰭT0+y0 _*tTk͏mh¬L Ҝj-hltʂR[gr"ɗҎW(&{4Yog7| @ N㫏<]?L*y]5WYCYxQTޓJvAWV}> TRX}͓(*+0uZ{(T+ɯAϘ)Pݜ,_ sG흇XV5>ygx&5 ^N=yMz%ȰAqXIX%_v8Z&>  ȴMSfbsLfix(fcŬ#jbcRIN@oy}3,XĿ7|iR_o %+Sa4ƸgVEȉ[0FQPlElD _N졒Ugh=D^=7v7Db5( W8]* ˨VPeAo*[{.*9DsNLjnw3Uwv X;ľc_.\hFI#Lc]Pޛ?"͘;M+ f =h}[IF5+?v.?r^?5 sJ`*ГˏqIn&ȅ11TOTkBVJ.b-!DVL5 1f܌MX@Ƥ, r@WO8*^8GD3'vW9acLb@wA\y̿wXfIB $QUU *,l`eJ*GrЎEA pDj&aVJ9J1ЗlIҜj--nXClBF1-o@u>^@SyCMR/ Boibm˲?/zP8^dA 4UV@Նs}8ʊJVj( 3!Ml?;YyƓ<6\|t[L||B @¤$O챬acg_/Tn}^].KnfW[-#\0$Jѣp$\d0ą8F|&.G|^1~-RtŻ+Ux9%t,= D rqcU|vJm%XŸ#"СLˆf9-uَ_E1+F?\|9n@Z~mWK.LAI"PbB CvQ=-Sm4;x R5ꤕ3gC)ŽaSuXv]лhѝPY0v+֪ x Nϱ!W[]YJ'ͷ6$`M֧ZpiutXX?OqΌT+C4ݯ17-S}'Ӽ',mʁ]7N3bc4넲q.HŖ':%yj%!R&:7Fa97 rn)O1C 1o/gs'Mr4˓7K_cG滚&PLsOTڱ!q4|ir8o|W8z K-xtƉz~L+,}z`@.aՖ:ɟV(Z]_>A8b03w>AB=j:B F YR1" ㎂A6eJ5,A*M,'8$t67,)H h5}wÏ `W( nF4C+ ^*]j\1͹NE4KS$3ĥy9jw_3FBATU^Ha$@X |BF*v[[FzPmLFN14JF*J )$: K" #m45H@dHnmAOM"p/b0 ]kl3D=3YZr#P#r$8Z$`ab-p:' xykGMs2&. 8ONj-Z"(vi$"?C=,>rCa f3%X'^qWo2̈́k4SmicC.㽥rB΃v0{yZlcv"YK<Gpn?rFގ3}޶p#izN49'uGGJ 7*;mT@WQ1tWЙ5-15?4P.;6R+]{vP՝.J^uCwRw#ǡse`RϿ~[Z%s`AZga<ߨjg  8qeS~>oY0: .3Z1:BFhX7>6)|0IFNgθYy;DN6qT`..L z~9MΌ\:~q [~hD1Ey~Wr >[HȻѝo^A}nL$R  l~TO|e^izbnaY4Svom~3oC umO{)}c/]u1'R3:LE;ԊoS,B!p1$=xkDPsԁ&IR&q&戩B&"qRII*If"2MR@y,cSf1-~[Њ4ucQ/y߰)6P%ɳ*P%ݛР %{Ic)J;C?؀L'U & ;:tnw92;HK"&pC)d.Lyf9'U9r32*<7?HEGJ% _5~~M^+=7&k?(0t>oqV 'bFs2.!"P'"-3b6S¹}{M^xauJhJWdQwi[noP>=jmvnM7O4?yT5 21c:J A7L-4*I2K@T&N,[BWWOb$&!XQKM:Lߤ^f tjL:!Y-rm]jI. Ij0q,6$$ xa;FE}Co7awj7!#}l0@eW@_ipdL+47rN,v98Smˀ8=IfIF)cX.>sMooU y8\/V_VߤWLsl|ҡ-{^s],Y_;6޿<"`}.ݸ]7j4T *Z)1OR&x$Q '2ϹbҎʌqL<\i%2M&4"#cw)JBB=kZ'5.:BVF+t+Y_{{ǂp<5FIBAk=A`+u :SJ;6=0E3gS6wb#T•eYZ.%9ʹ>0*.pئ,FsЧj.D1c qvndFikgR TPoZD, l5) 15~bd;(m)~!RhDMiҭ)0\835"rirNpW{֗!B6PJ0}B֨r1l{K#挹ӒwC=WVHC hu_7~Oko7o:}@ml}w_Qfr=VΞ= .tuOouʫ'cCTiQ&5O8WӫIޗUm젺]W]/?\0@}K*:e@IGGuvUxL59itRQmQ{| N4u ׇe# e c4rl$pBWV=ZuupXvPNmTU nkz>zڄ2 ӞFTڜj04<%#١Z]@ni{?hݗz&TfsUw;*Ω.hP C%E٢o]zpbv[ՠ q/SŷԒY`J7./-E:J2埣AU0%,d|fw-BuC L]b6]=GۯCX 6<%:JPc1"\'N =tJedوwnWv]ёg9Ū ' 8 i BeWsTIGb J58? a물U4H2Qi4-r)'jNe>-,V7rP__ e?^[\3>vїN8@[/LӮhx!R|byg[B|Zz\CPs 0|KK aasB[yn7|!@5ԽtS8Va &1%w JuޅK-5tVfoėg{zX,~z\((%GUч_{yW\PPr쇟oR3 At /UOqm*[ tפyQWpʒ{cKŇ.׿/{|MSaί7?^\OB7oT>cw<]Dwx6S g ǢLڝۯߦl'Nm*'.9[WTY B޸)Fݔ *y8 t>vFSňFG!,䍛h/,nnZ9n}11h3q :lѭm y&٦fe!rj"JCvBV2 KֱKinSŤ)3Hfr5Y*ɩԉJ'`|Qp*QNJMrW2R187PBڠh!ڠ c hL&bhmP4$8\ !@z4Rɘ)(B[KnM9\^s%'!诋hE22F9Y_gc!x[M c=wCÙf/.+n%{\iBOր"q\ b ?-ҔZ"KiH gU6fWI =gΌJL˧S 5=U*(Äi4 |W`Dv:5r" 7d~+  \Ùpi5ł01@(RDUnDNΔ 25ɝ S)HcTmJ$0EBpL &u?5AÅ0@ L:H6Jȿbaƺ,qR%!xF4IK\\j t:&W`. 9|pl8k_3×2|Y>P$P1Jp Z/ Dy~ c1/K=Br8Q8v8Jxo|B$/=c0ۡ ~bDif>G2yJ{S$Alm/_ /hP[=ܭ^գ9U99AX#kgsёg<+⃇=̺PAFO>5OGy˯~%/C;imw˯a4Xb,b~gGZ7Y9p-Tvf=oQ|`M](CADl('\f9.*$g4@8 9O+|Jt+[2q<(j;+Ҏd$4#t)MrBS|ɴЄ:a"ǵ7J} W^N9H mȘ0 4`gE%4&>qyV>U1.{({i 5J"a?OF*(kFoM7d1B])c /ۧcY:{,+y_|=hW6C7ZzDZܦG^-x0ٶVzF㇋?zpuyI WZ}YMrXEZ=Μ/t+~-./zI%=koI}Ѽ큎[?Z}Kٝε@4h id"iʵ;xoF:9v ](ZI^4?{׶Ʈc%o/ HS3?pTɲ-)qv[N$n]]|nMfO"Pr;^%nײ17C1gwOc+IziQiqpn}tGo~T7S8(h %Qn.Ȕׇ|IC?$RgݎI tq^BKlj*paxsWS^Dx<wR(Nqwfa||J*nϧGҍЧauӾ͚lΚ400N05.ٞVVb2zN seP5cWpͱՆx wڜAX ɷ3Δ?^-Ԗ\7W\: J@??ikC!xj/im5,gW1?\BrXImH]]UrWU%wJ~!.XVt9)Qz]RȋtȜ| 0$Ԗ6QJs,!J_/Fq6p2OCOO- Y)vqWiؾȝƷ:3{/vw7ud2))b\@qd;r&sb}kYKսD~}]CDHFhթzOB!Z#z/h%OO#O[o?3-D6Հlf(UAj W?1l 9knQzz}?wR~~5ht?ng&Fv{B8hwɤ\Q 5!&s-쟪9J΄Ös(lu=8YfM}HRe==[u|v/pv/Wݪm:K%ͧn1RIFA㗴/4ƮwZNj2oޱ15z;:@NjVM1#4NZFyҠM' s%Z6X)d iz6]=-bmZevBZyP.6 tڢnzp)hd:J+y(<ρ9x#L&pY(o5M{mͮp2%YD+EW2^{6I&:f$~u~5cXաDxi#[ y}sɤ0YXy"S xLz}8ЩS[ԑqݭ+Z-Ō+=f{tZyh_9 B2$4eiwK%CXGw1_n)IZ0ޑ>$+S5i2^.yW9,Q4oV4[ӫk~*Zk5jSC\^PLȃ<bZ ͸rQUqLDckQ [Ou혌Nx@gY>f@^)I:(zsFl(m'ŶNFP7'lW5hY]wx,@PM AXJ16M?C%e+]v=͗~wf.YY\wpoS?RdL=ZyJ_ LMOy~鴬~-,:ʌ*+" "+ hh~jE#D1Z!Oc{%vj)hHM,H䆏 Q՛zQEpP_՗ן]\YB- 2ړLBٍ?IJuvWЙbvdZ3JL~r(>#jjrB?b [rvJF@)*N{ ;T*2lSb@ :}]v--XΎy Xv2{c3 Ūn<^UcQx]JYDI+ ZE΂ʊd*J8l<؁ Ҳɯ|q]hg %%Ca`읖{u0ӏ[NQNj9^og5}f2 J@-L_iL0ey/Qb9󟆢udVBN@b^K2g)hqY;}-48ZJtŮ초+O,$KT{41wˣ $&~yTRd6")>Ś fθ9K W~!u;2TFydcL!0IH^6*,] *ŷZ/q"YAm܅ZeJVCCy+}d^JG> uezaƞY[CjJo Upt %gbb=+{au*[ܓ]<6Bc)ǒͧwj,K3CzFiKXYJY5RmFMZaǎELvx;'(U82 iMD.EXĢ\X84J;)n\UPYN25\N(P_RM*C8,gy/tlN'rml4)´I|5=2 b51i~Pw_lem.0*Psl!DEv)SDOoY"XnBI5 /ǫULi}8hݺMi4-)_(:"d($Z(Ro*b8Pi0Ms򖗉iC%V/CSCo/)iRoSF{~b(Gi;@i=y$`CgbІ)5Ɣȭ-r)qEP|"søSHI!CUcPsI˓*x'l4)[&4%aZcV'8Ly4؉pĄ+B˱RÃvd| u(6޻.70g` "XfFi azTO`=Du |6Jx,X6LU΁l[t|%'B3,Hz@Btf VFL~Zp2ڴ`V$Ӵgc!05Pпڱ_Ytۘfaq%G1pΜ<_I"L %Դ}22cT:ALTJg W p@E>7Pb~ad7wf:@aRL KҞ ү]3`$yqv +_:V@U, GAy>n4nWlچcM+q]eeG1NJQ6,^̐ysOws3{V7 =hCeurQYɕMatDz8l\-E ,"XBeuSV6X2?>k3h`jP5P^+3;Y 砮GRRVXdJá9Ji#3ޕ5qcӤ&/CJݙ8y4 ѦH,{NE5ŦD%ʃ#nhgG%E\;KMSHkFNaT}A-pD[ VKFᐕ.GҲ\93r*H7ٍBurʉájhN;yOh~QHU5[fk&4}m0TiApӔ&%AJd3xsaQ]jamΜdJ 7^M(%GgP} 8 e2[*K!3+;^?x6s9'T#8L?d-Y.rDD% #޻@APqX!`ܵ͊tٻ&Q tJ%%L %c}<^3='79!m_CU_G1|<-4 :YFEUbgh2z>57@/#kld%/-#kl' q|n"Ki,4h6S@bɴws?mk² h9j3.0QGGϡUC:(g,CLg4y&}VAXjd迭(r __v֛ZW /~Q1wke;aKoz1< P X!`vV9?W֔6ɪ,$,XsX-Ad<~ߙ|FꨓÄ}80ZZn>f'ghy]ss:;aKξ^Ȩ]ܫ,k}8z[lX)h;[zt{Kp.L w#Nýمa#uTܛ$8OA'V:j ̛E+$S&|'\%ShF~ n>듛 vˢ/r[NyE.)Ci!:l+*=seA[m֦ܪa"yj!aM>9k),{YsxW;~hYܜoy稥6RFHZ K&͊Gd)R+K%n^'L=LxCx7@٩.H ,EX-qU;F<5.e+6HϚ Yw68s 9\wŞ6鄕ym0* 5tt='XCN`h|f&b.Ov=D6dS䌛|V0 V+=0 YK:/~!R-1`zc5R#6yV)/`Y# gb$6)E5B\.Y(yHԧe(,U%H)!2AT! hHrJ[t\|\֩RjpC#<%W,YOn|+fv_A/n^ $H&&\Dx^ñDFNˠ  { ]ʘ{{AͫޕΨ`"oPeذ9"%3(ixU{stFkR(!! bµE>Zm${W2kmԎE.L!CJ2hJ#ei)#EvQ?2ca}kEy,* _Db50i'U*sk;,¿H,ːAEqT$pԫDՈ' zP̈́y3 72[_٨ L!I`, ӑ9gܷEҺn5K2%/O/ YYjzp_ΞB@ۢ徿cݳodE*Cf,髋LA%B)~K(E-N8c4:W#ҽ$8gvVDtD,%)"~BR/wZגZ# ȳ 0aD[F  1!v8MrK+[ f/.42f"S55p,Y'&"4E %qiR iFwOgRHҙ Uu—iF u,vOXZc?\&Ӻ_z+!!Ont^xHCVa➕ޜz+,{{6:6K6:cSHq1`AV W7B1V`6[)0 x0nsV/:ۣd8ȏz&"fX(* W{϶|'7 Q쵨hVK#]Xq}g^:wV.j\M.A e"`v[. 2(݅UXy}-JMӎ!2#`׺<`Ůǵ8*ޓԧaxyUg߉~p6(WƸ;35&x8 9u2e?v+t]WT "q`sN.SƽҎ0V%Yu1Q]aSŘ&Kt{?}%TW&;awT3ؑ2QBibg|ẆXǷou|&}3NGαR|S ~w u, < 8PK+S*9wgJ$'ӫ!Ψx N b7\`c!a3ippASjLuwǼA4hxFȪK)6q5K=b)pcۺkÜ{H^[ƎvsAs_f=Iky+$\5}\,cR?cKHF@ ^_' s4aƢwG1$)27h^sҼ`z|F8)]^K;߽!DsL˔[{L~p~k ӿOdd-nIZY;>xCrrsRns$\*:c'V|w!ZVgQ?s5bpO47M"Yj:/5j>,kq zcggaToA44sQ{9ĥ?%a~cr3˹ WiZι{sD4Mk!qIjؖn[7{QSX]yDuv}6 fvlhАSP#[M1#"vۑ*v{,b`7Zn-h_\EWuJ6ꑬje{R1UWjT7Q{fp<:};xbeQT_//=dųo߰HYJ۠]{ =W]W,Qdj=N!]&ߢ4*<"{AH46dcN9x-X%8Z?fNF㷋BFQW$ld˂& B!~6jg_;d1Gd4^~]OyKip_d4[N?}PK2?xzukGH> _B`xH%I,Gnx ʃae}r A T/g(]~4."ip[M ,ZDVv%] w%Pk2q͗ORx0y x`$&K*Nj mty=4loY[Z/va׵/r487^ɛ 7JUET4`TgoCo iJ %@:?{O۸~L(^'n榑3/ R ~OQM˲(R6V?$LVg:x6nҴ99 XNXņ%F*f < vYKHf,/2eq3T}) lXƱC;Ca}` }Zg\ȄB  >p=Q[aj߄910aSxй >+YFx5MQ1ֽ*0+l!|bp1ΓȄvBQrc޲apmVN8T9ja7"yJ|/A )e&MJK2 {-"18y.gZlq6J:e" d (Z@63QB{!Ձy'TQM⬋>zR G Ztl6l@NGG.ξ9ch)r$ c;: 'Tq?C$:ԏƹ{H ŅџZcdM;cS fc' 3T\i&&hNKXImt;S6;eN\e߉ Cq+B O7+$Q+ aQ`}B@l+Jdcxw<ǵa ;ɲvRu\*/e ! "]N>L*pA3A,dE Hȅ/kJщӻǠ>aK=?KR)I}U΅`tf;Y Xguk99<2o(+)gZ!z70>hɘ-jf!^2T]a8M|"^>2#ΤM Iቨ>HFv-P[}e3Jו17p3Ftd"$7DF@',cPCi`2Wih%mt 1VRj <7EQ s3jzEc E*BL2 WAdvM'}d@$ F9!&QcBHB[%sBhNֈiA"0@:;!8 gޔ NhK a읕gx`,GG<!fPHS `yۈv*L!>g8jبҰLˑeA]j.hRH>p ƄTͬxƠ,nZ~QΨدζ{{X؆|)':&;Y V$dJnv (gOPL($vis8s^ׄTn}f/eӈz^{l6JZse-\k,L , Wx:,hsBU"*WPo-SJG(i,l1I HgPzo^}oXMbꑭ=i 7Z͖kW]Ɂ?0||uӲ}~*Ǭ-y순ZN^C?cX,Eo/|c+vp|*\cfқ-R}c;T?h:BWbv8pB[rQ] i"4#w XltRe~Re3 Y}>o{nޟ?~+EQFC׬XL)? AAAAU"W/8Ӟ* jer8B1J RBhI ZHVJNzk0rw{M:íŬ'@g\ Eor,Z^ ]פ%2SLB܎<}ߞIe*9ޥ蚵࠭1ϐx,K+*5ǤC,- rWcOIhln+YϣoΎNQ܄͝j1)jB@Na2a]؄iZɣ JMK{UV}PU]Ɋk,z;+;9 7B| dM9b ~n}YLQ5b~?鄹7^zIBN$!.]Uj,gU -+Q(pL)9%vpL1eFRkR4 {jhh8Enh%YB(e5J )ŗ3C[qWW]ۀ\Q>W v%c]Q*#W^Fo[Y ~wS&HX¤sv~vQ >jafK^#*5rrzSaY p:׋æAo.Oi(7g"Gާj1J(QH _:Б!1~}Ol.滤V{}gd c*ZLG-V<љ);uy}绸ۜEZ^/:xRt/*e[ 1v:'Ǎ X.5{)u^ STdo.;՗Q? u%Im2?aXI֍C 9LtݡpGO6'}$Hs?]%nwɊfjI״zmbx-dnE[msq3q*rVco?瞜< 8TWx+Ci|ջǛE|YG@ZS Z׺{jUz YXA َčS@R@5|3XJv$nξ= 7Xz)[TuثcbO'o˟=[=؋.yjaqZz 6 ǫG^|j.̽<^TsbqN/{ 댋{:˪袡x;Aufw;^z[G4d㛳MyA5GN5~lc_')k)M+ d[]% "ygjc(% Eٜ 3-kPDRZ{1ס6Fb(7&$pAօ L [hi\f 3Y``a Dh{ AA!,(#hi,6OKY*tcԡ4-huxԼ>1qNzĎog ArpIqd8T.b,mG˖,>Zbk HIS fe8S 5sjz:N( a68OQRH !(ENQ8' o#tXmZL/A+H2HGmjFZ A+ͅ_oM432GixǶ N.;l2۴ui *i%ɪ{a\P%5:KrIE'brUaXU&=Z/zz8^oNeY-(j1OWL~P7<䂾Vj1 ?)߉W&) ί&6'm7Zz59uL$Q24}XRҞTHDUґ[V!t7`η8sÔtt<yn%<+ :%=@01 N B娟K&k%2W9z7 39> ^*5MD 'i#dw:;5MG;jZ} }hg"$;Gz|A+6vvl~r;޲Ֆ{ 8"h7A9y~pkq6/A<|B@8E 咶{U%q%weqr3Er7yܮ/Wo^[[fweZX}=v4$IDoA|MBNSz$p.oM'a$| YeQK6D쀥T|KB)UY=~o%&"SdAN H-Gڛ`' sv+8ե*p7S}x I۸FUQůC9u{,Ay.ՆLꄊ7iMNUYI UƔ|NpM1n<޼eY ڀIfaf !!E\l̬ƌ:Iሓ\T!X5qn TJm\89lH!Nؾ>Iat$.K#a qDD-^`n*hnaC\4ݱbZs\B! Y^b|6jµKiS;HmUoǫW1à\#KI m41\Ju A뒽L'0@86`9d@R-G%gp.7?j^x;3G74,2Qy5g7C&EE\ M1SSbPpPG4WiQBÅ\n {{bpٝO*_wbdHa,|YjxK"yKFQOR.xoq9',a{#"j Gi踜"\!n{Z+D5˩E\N3D)0!h٭_Xܶ3d1GtZ֫ozӘ9NcdҸ֪w >4k tlZܽ?[ղSHnbFJAeOS";f};]fD{xp!' 6MO8Ccg6>]/z1#&~٫aCO)~C:/b Il#Ve 9f 0\^s PC8#  3(ی;HK[%sC i--\C##і4sXX0p(gxUgo9e:/p,;a/s"^}{p{9[I.9a0|}^p a;:T#wt;_0ǻ;Y mWQ[Sno(o{]0xgPͳiHjmdpP}5\߰||] Y}3E#PՆlGڈȣv.'6]>/J .UU0b'ߝwvx3,>N*px,B<`UOΚ 91v03Œ5mI/k~)zY\v"p ( ^Mbj)z@*Չo̗awY/:K/dOrQ4<`bQٔQ\JKdH- VS;k[ZEsJup1M@9Vk,]c2W>-M$R-q]q"Qߠ$rK#TÂ-J j{#Dc]NkՌ 7b1ΙBO[FykVJ"+֌q{V-A[q;"h+񾠭]lk𙡵Hn J޷ff  !h1KkO)/ tC| A3.~8nG A27?R x-, IQ[ЈaiAig\AZ!\%WQU2)f?3, q9'%Y!8wH"i,b)$HSj!/:Pպwm4:J5Qv#9 T\Z0-ǹ R/ \(U `XGǐ2aք˓^AʢڋϹWiQzEβ`feӫT < y|52@ag67Es $\Y?&x__TM.:*[8t^Y+og]L9? 4ڪU>L44KvoP6YIZIQנ:O\DSd\ m&kO'n<1h-Xօ|"$SB9$"[*bD'u&G,)cidBc[hk 8Rb".Xm71gXܝr3g!> 7V^ǽN\'~f/nyW2dz>%N>$\;|5qݸpd]5֏w 4BsXD ab`}rj,l~]Rj q;ef%.<F3$hΊLf<m=p1>&4XH,H^>Z1Y*I_?`m*QA;var/home/core/zuul-output/logs/kubelet.log0000644000000000000000005471257415146025156017721 0ustar rootrootFeb 20 06:40:42 crc systemd[1]: Starting Kubernetes Kubelet... Feb 20 06:40:42 crc restorecon[4460]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:42 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 20 06:40:43 crc restorecon[4460]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 20 06:40:43 crc kubenswrapper[4492]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 06:40:43 crc kubenswrapper[4492]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 20 06:40:43 crc kubenswrapper[4492]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 06:40:43 crc kubenswrapper[4492]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 06:40:43 crc kubenswrapper[4492]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 20 06:40:43 crc kubenswrapper[4492]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.421790 4492 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426256 4492 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426278 4492 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426283 4492 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426287 4492 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426290 4492 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426294 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426298 4492 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426302 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426305 4492 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426309 4492 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426312 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426315 4492 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426319 4492 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426322 4492 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426325 4492 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426329 4492 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426332 4492 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426335 4492 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426338 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426341 4492 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426344 4492 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426348 4492 feature_gate.go:330] unrecognized feature gate: Example Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426352 4492 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426355 4492 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426358 4492 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426363 4492 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426368 4492 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426372 4492 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426376 4492 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426379 4492 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426382 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426386 4492 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426391 4492 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426401 4492 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426405 4492 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426409 4492 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426412 4492 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426416 4492 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426419 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426423 4492 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426426 4492 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426429 4492 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426433 4492 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426436 4492 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426439 4492 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426443 4492 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426448 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426452 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426455 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426459 4492 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426462 4492 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426465 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426468 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426487 4492 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426490 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426493 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426496 4492 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426500 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426503 4492 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426506 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426510 4492 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426514 4492 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426517 4492 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426520 4492 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426524 4492 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426527 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426530 4492 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426535 4492 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426539 4492 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426543 4492 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.426547 4492 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.426954 4492 flags.go:64] FLAG: --address="0.0.0.0" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.426969 4492 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.426978 4492 flags.go:64] FLAG: --anonymous-auth="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.426984 4492 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.426990 4492 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.426994 4492 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427000 4492 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427005 4492 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427011 4492 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427015 4492 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427019 4492 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427023 4492 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427027 4492 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427031 4492 flags.go:64] FLAG: --cgroup-root="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427035 4492 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427039 4492 flags.go:64] FLAG: --client-ca-file="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427042 4492 flags.go:64] FLAG: --cloud-config="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427046 4492 flags.go:64] FLAG: --cloud-provider="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427050 4492 flags.go:64] FLAG: --cluster-dns="[]" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427057 4492 flags.go:64] FLAG: --cluster-domain="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427061 4492 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427065 4492 flags.go:64] FLAG: --config-dir="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427076 4492 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427080 4492 flags.go:64] FLAG: --container-log-max-files="5" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427092 4492 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427096 4492 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427099 4492 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427103 4492 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427107 4492 flags.go:64] FLAG: --contention-profiling="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427111 4492 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427115 4492 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427121 4492 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427125 4492 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427131 4492 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427135 4492 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427139 4492 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427143 4492 flags.go:64] FLAG: --enable-load-reader="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427147 4492 flags.go:64] FLAG: --enable-server="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427150 4492 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427156 4492 flags.go:64] FLAG: --event-burst="100" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427160 4492 flags.go:64] FLAG: --event-qps="50" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427164 4492 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427168 4492 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427172 4492 flags.go:64] FLAG: --eviction-hard="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427177 4492 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427182 4492 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427186 4492 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427190 4492 flags.go:64] FLAG: --eviction-soft="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427194 4492 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427198 4492 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427204 4492 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427208 4492 flags.go:64] FLAG: --experimental-mounter-path="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427212 4492 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427216 4492 flags.go:64] FLAG: --fail-swap-on="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427220 4492 flags.go:64] FLAG: --feature-gates="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427224 4492 flags.go:64] FLAG: --file-check-frequency="20s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427229 4492 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427233 4492 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427237 4492 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427241 4492 flags.go:64] FLAG: --healthz-port="10248" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427245 4492 flags.go:64] FLAG: --help="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427250 4492 flags.go:64] FLAG: --hostname-override="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427254 4492 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427258 4492 flags.go:64] FLAG: --http-check-frequency="20s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427262 4492 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427266 4492 flags.go:64] FLAG: --image-credential-provider-config="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427272 4492 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427276 4492 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427280 4492 flags.go:64] FLAG: --image-service-endpoint="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427284 4492 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427287 4492 flags.go:64] FLAG: --kube-api-burst="100" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427292 4492 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427296 4492 flags.go:64] FLAG: --kube-api-qps="50" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427300 4492 flags.go:64] FLAG: --kube-reserved="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427304 4492 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427307 4492 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427312 4492 flags.go:64] FLAG: --kubelet-cgroups="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427316 4492 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427320 4492 flags.go:64] FLAG: --lock-file="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427324 4492 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427328 4492 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427332 4492 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427344 4492 flags.go:64] FLAG: --log-json-split-stream="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427348 4492 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427352 4492 flags.go:64] FLAG: --log-text-split-stream="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427357 4492 flags.go:64] FLAG: --logging-format="text" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427360 4492 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427365 4492 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427369 4492 flags.go:64] FLAG: --manifest-url="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427373 4492 flags.go:64] FLAG: --manifest-url-header="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427379 4492 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427383 4492 flags.go:64] FLAG: --max-open-files="1000000" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427389 4492 flags.go:64] FLAG: --max-pods="110" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427394 4492 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427398 4492 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427402 4492 flags.go:64] FLAG: --memory-manager-policy="None" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427406 4492 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427410 4492 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427414 4492 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427419 4492 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427430 4492 flags.go:64] FLAG: --node-status-max-images="50" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427434 4492 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427438 4492 flags.go:64] FLAG: --oom-score-adj="-999" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427443 4492 flags.go:64] FLAG: --pod-cidr="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427446 4492 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427453 4492 flags.go:64] FLAG: --pod-manifest-path="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427457 4492 flags.go:64] FLAG: --pod-max-pids="-1" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427461 4492 flags.go:64] FLAG: --pods-per-core="0" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427465 4492 flags.go:64] FLAG: --port="10250" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427469 4492 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427490 4492 flags.go:64] FLAG: --provider-id="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427494 4492 flags.go:64] FLAG: --qos-reserved="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427499 4492 flags.go:64] FLAG: --read-only-port="10255" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427503 4492 flags.go:64] FLAG: --register-node="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427508 4492 flags.go:64] FLAG: --register-schedulable="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427511 4492 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427518 4492 flags.go:64] FLAG: --registry-burst="10" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427522 4492 flags.go:64] FLAG: --registry-qps="5" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427526 4492 flags.go:64] FLAG: --reserved-cpus="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427530 4492 flags.go:64] FLAG: --reserved-memory="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427535 4492 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427539 4492 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427543 4492 flags.go:64] FLAG: --rotate-certificates="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427547 4492 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427550 4492 flags.go:64] FLAG: --runonce="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427554 4492 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427558 4492 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427562 4492 flags.go:64] FLAG: --seccomp-default="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427566 4492 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427570 4492 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427573 4492 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427577 4492 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427582 4492 flags.go:64] FLAG: --storage-driver-password="root" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427586 4492 flags.go:64] FLAG: --storage-driver-secure="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427590 4492 flags.go:64] FLAG: --storage-driver-table="stats" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427594 4492 flags.go:64] FLAG: --storage-driver-user="root" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427597 4492 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427601 4492 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427605 4492 flags.go:64] FLAG: --system-cgroups="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427608 4492 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427614 4492 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427617 4492 flags.go:64] FLAG: --tls-cert-file="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427621 4492 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427627 4492 flags.go:64] FLAG: --tls-min-version="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427631 4492 flags.go:64] FLAG: --tls-private-key-file="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427635 4492 flags.go:64] FLAG: --topology-manager-policy="none" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427639 4492 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427643 4492 flags.go:64] FLAG: --topology-manager-scope="container" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427647 4492 flags.go:64] FLAG: --v="2" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427652 4492 flags.go:64] FLAG: --version="false" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427658 4492 flags.go:64] FLAG: --vmodule="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427664 4492 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.427669 4492 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427793 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427798 4492 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427802 4492 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427805 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427809 4492 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427812 4492 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427816 4492 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427819 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427822 4492 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427825 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427828 4492 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427831 4492 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427834 4492 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427838 4492 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427843 4492 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427847 4492 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427850 4492 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427854 4492 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427857 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427860 4492 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427863 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427866 4492 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427869 4492 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427873 4492 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427876 4492 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427879 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427882 4492 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427885 4492 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427889 4492 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427893 4492 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427897 4492 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427900 4492 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427903 4492 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427908 4492 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427912 4492 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427915 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427919 4492 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427922 4492 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427926 4492 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427929 4492 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427932 4492 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427935 4492 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427938 4492 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427941 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427944 4492 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427948 4492 feature_gate.go:330] unrecognized feature gate: Example Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427951 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427954 4492 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427957 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427962 4492 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427965 4492 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427969 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427972 4492 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427975 4492 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427978 4492 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427982 4492 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427985 4492 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427988 4492 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427991 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427995 4492 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.427998 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428001 4492 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428004 4492 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428007 4492 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428010 4492 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428014 4492 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428017 4492 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428020 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428022 4492 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428028 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.428031 4492 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.428038 4492 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.435513 4492 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.435544 4492 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435623 4492 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435635 4492 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435640 4492 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435645 4492 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435649 4492 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435654 4492 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435660 4492 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435664 4492 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435668 4492 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435672 4492 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435676 4492 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435679 4492 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435683 4492 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435687 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435690 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435693 4492 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435697 4492 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435700 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435704 4492 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435708 4492 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435713 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435717 4492 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435721 4492 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435725 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435729 4492 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435733 4492 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435737 4492 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435740 4492 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435744 4492 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435748 4492 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435751 4492 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435756 4492 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435759 4492 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435762 4492 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435766 4492 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435769 4492 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435772 4492 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435777 4492 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435781 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435785 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435788 4492 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435792 4492 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435796 4492 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435799 4492 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435802 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435806 4492 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435809 4492 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435812 4492 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435816 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435819 4492 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435823 4492 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435827 4492 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435830 4492 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435834 4492 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435837 4492 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435840 4492 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435844 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435848 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435852 4492 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435855 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435858 4492 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435861 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435864 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435869 4492 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435873 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435878 4492 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435881 4492 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435885 4492 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435888 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435891 4492 feature_gate.go:330] unrecognized feature gate: Example Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.435895 4492 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.435902 4492 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436059 4492 feature_gate.go:330] unrecognized feature gate: Example Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436064 4492 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436075 4492 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436079 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436082 4492 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436085 4492 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436089 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436093 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436097 4492 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436100 4492 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436103 4492 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436107 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436111 4492 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436115 4492 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436118 4492 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436121 4492 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436127 4492 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436133 4492 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436138 4492 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436142 4492 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436146 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436149 4492 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436152 4492 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436156 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436159 4492 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436162 4492 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436165 4492 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436169 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436172 4492 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436176 4492 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436179 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436183 4492 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436186 4492 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436189 4492 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436193 4492 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436196 4492 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436200 4492 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436203 4492 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436206 4492 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436209 4492 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436212 4492 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436216 4492 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436219 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436222 4492 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436225 4492 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436230 4492 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436234 4492 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436238 4492 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436242 4492 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436245 4492 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436249 4492 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436252 4492 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436256 4492 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436260 4492 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436264 4492 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436267 4492 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436270 4492 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436273 4492 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436277 4492 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436280 4492 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436283 4492 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436286 4492 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436289 4492 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436293 4492 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436296 4492 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436300 4492 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436303 4492 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436306 4492 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436310 4492 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436313 4492 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.436316 4492 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.436321 4492 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.436496 4492 server.go:940] "Client rotation is on, will bootstrap in background" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.439223 4492 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.439293 4492 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.440176 4492 server.go:997] "Starting client certificate rotation" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.440201 4492 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.440804 4492 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-15 09:59:10.976164575 +0000 UTC Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.440870 4492 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.450342 4492 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.452044 4492 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.453220 4492 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.466216 4492 log.go:25] "Validated CRI v1 runtime API" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.483614 4492 log.go:25] "Validated CRI v1 image API" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.485046 4492 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.488038 4492 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-20-06-37-09-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.488060 4492 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm:{mountpoint:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm major:0 minor:42 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:50 fsType:tmpfs blockSize:0} overlay_0-43:{mountpoint:/var/lib/containers/storage/overlay/94b752e0a51c0134b00ddef6dc7a933a9d7c1d9bdc88a18dae4192a0d557d623/merged major:0 minor:43 fsType:overlay blockSize:0}] Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.496772 4492 manager.go:217] Machine: {Timestamp:2026-02-20 06:40:43.495751442 +0000 UTC m=+0.267040441 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2445404 MemoryCapacity:25199484928 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:b34febcf-795b-4446-bb24-1edd1f924aef BootID:cafa21b1-6432-4073-bf29-13d2dbbad80c Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599742464 Type:vfs Inodes:3076109 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm DeviceMajor:0 DeviceMinor:42 Capacity:65536000 Type:vfs Inodes:3076109 HasInodes:true} {Device:overlay_0-43 DeviceMajor:0 DeviceMinor:43 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:50 Capacity:1073741824 Type:vfs Inodes:3076109 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7c:e8:ab Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:enp3s0 MacAddress:fa:16:3e:7c:e8:ab Speed:-1 Mtu:1500} {Name:enp7s0 MacAddress:fa:16:3e:75:72:c0 Speed:-1 Mtu:1440} {Name:enp7s0.20 MacAddress:52:54:00:6f:63:ae Speed:-1 Mtu:1436} {Name:enp7s0.21 MacAddress:52:54:00:99:27:cb Speed:-1 Mtu:1436} {Name:enp7s0.22 MacAddress:52:54:00:d7:79:eb Speed:-1 Mtu:1436} {Name:eth10 MacAddress:1a:d8:3f:9b:73:76 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:52:41:d7:24:cb:60 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199484928 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:65536 Type:Data Level:1} {Id:0 Size:65536 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:65536 Type:Data Level:1} {Id:1 Size:65536 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:65536 Type:Data Level:1} {Id:2 Size:65536 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:65536 Type:Data Level:1} {Id:3 Size:65536 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:65536 Type:Data Level:1} {Id:4 Size:65536 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:65536 Type:Data Level:1} {Id:5 Size:65536 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:65536 Type:Data Level:1} {Id:6 Size:65536 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:65536 Type:Data Level:1} {Id:7 Size:65536 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.496932 4492 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.497020 4492 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.497892 4492 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498047 4492 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498088 4492 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498253 4492 topology_manager.go:138] "Creating topology manager with none policy" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498262 4492 container_manager_linux.go:303] "Creating device plugin manager" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498550 4492 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498683 4492 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498806 4492 state_mem.go:36] "Initialized new in-memory state store" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.498874 4492 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.500816 4492 kubelet.go:418] "Attempting to sync node with API server" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.500834 4492 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.500879 4492 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.500890 4492 kubelet.go:324] "Adding apiserver pod source" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.500901 4492 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.503095 4492 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.504166 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.504313 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.504209 4492 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.504276 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.504541 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.505972 4492 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.506971 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507048 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507105 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507149 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507198 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507249 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507309 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507358 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507408 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507453 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507536 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507591 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.507654 4492 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.508118 4492 server.go:1280] "Started kubelet" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.509050 4492 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.509204 4492 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 20 06:40:43 crc systemd[1]: Started Kubernetes Kubelet. Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.509366 4492 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.509909 4492 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.511766 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.511799 4492 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.512207 4492 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.512222 4492 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.514201 4492 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.512887 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.514658 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.513642 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 01:36:04.186658938 +0000 UTC Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.512357 4492 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.513452 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="200ms" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.520360 4492 factory.go:55] Registering systemd factory Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.520385 4492 factory.go:221] Registration of the systemd container factory successfully Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.520726 4492 factory.go:153] Registering CRI-O factory Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.520743 4492 factory.go:221] Registration of the crio container factory successfully Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.520797 4492 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.520816 4492 factory.go:103] Registering Raw factory Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.520916 4492 manager.go:1196] Started watching for new ooms in manager Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.520539 4492 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 192.168.26.25:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1895e134602a4089 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 06:40:43.508088969 +0000 UTC m=+0.279377946,LastTimestamp:2026-02-20 06:40:43.508088969 +0000 UTC m=+0.279377946,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.521344 4492 server.go:460] "Adding debug handlers to kubelet server" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.521460 4492 manager.go:319] Starting recovery of all containers Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529553 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529594 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529606 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529616 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529627 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529636 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529645 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529656 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529670 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529681 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529694 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529707 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529718 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529730 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529741 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529749 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529759 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529771 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529783 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529791 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529799 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529808 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529817 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529826 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529836 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529846 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529858 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529871 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529879 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529887 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529898 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529910 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529921 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529948 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529958 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529966 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529974 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529982 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.529994 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530006 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530015 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530026 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530037 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530044 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530053 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530063 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530079 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530090 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530102 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530113 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530125 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530135 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530151 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530163 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530174 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530184 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530197 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530209 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530219 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530229 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530240 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530250 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530261 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530272 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530284 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530296 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530305 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530316 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530329 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530341 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530350 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530360 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530368 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530378 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530386 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530397 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530407 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530416 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530426 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530435 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530448 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530458 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530468 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530507 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530518 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530529 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530540 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530557 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530567 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530576 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530601 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530614 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530624 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530635 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530644 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530653 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530667 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530677 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530690 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530700 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530711 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530721 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530731 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530740 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530754 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530763 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530775 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530786 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530799 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530812 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530823 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530834 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530844 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530855 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530866 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530875 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530886 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530897 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530906 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530916 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530930 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530943 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530954 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530962 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.530972 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531225 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531237 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531247 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531260 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531270 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531279 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531289 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531299 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531307 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531314 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531325 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531334 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531343 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531355 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531364 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531375 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531385 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531393 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531403 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531414 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531425 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531434 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531443 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531454 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531462 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531482 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531492 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531502 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531510 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531521 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531529 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531537 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531544 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531555 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531566 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531614 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531623 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531631 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531640 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531650 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531658 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531668 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531676 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531686 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531693 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531703 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531714 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531722 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531731 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531740 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531753 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531764 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531773 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531782 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531791 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531800 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531810 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531822 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531830 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531839 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531847 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531863 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531870 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531880 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531890 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531898 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531907 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531917 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531925 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531934 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531942 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531951 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531959 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531966 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531974 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531984 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.531994 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.532002 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.532010 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.532019 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.534655 4492 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.534815 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.534826 4492 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.534837 4492 reconstruct.go:97] "Volume reconstruction finished" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.534844 4492 reconciler.go:26] "Reconciler: start to sync state" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.538843 4492 manager.go:324] Recovery completed Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.553841 4492 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.553881 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.554839 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.554866 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.554877 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.555643 4492 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.555676 4492 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.555702 4492 kubelet.go:2335] "Starting kubelet main sync loop" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.555742 4492 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.556100 4492 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.556118 4492 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.556134 4492 state_mem.go:36] "Initialized new in-memory state store" Feb 20 06:40:43 crc kubenswrapper[4492]: W0220 06:40:43.556693 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.556757 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.560918 4492 policy_none.go:49] "None policy: Start" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.561658 4492 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.561683 4492 state_mem.go:35] "Initializing new in-memory state store" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.598514 4492 manager.go:334] "Starting Device Plugin manager" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.598550 4492 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.598561 4492 server.go:79] "Starting device plugin registration server" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.598926 4492 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.598942 4492 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.599102 4492 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.599379 4492 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.599388 4492 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.605020 4492 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.656028 4492 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.656105 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.656886 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.656916 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.656926 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.657038 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.657708 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.657758 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.657768 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.657918 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.657942 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658108 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658137 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.657951 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658695 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658711 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658722 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658854 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658906 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658938 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.658947 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659251 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659283 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659291 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659365 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659378 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659385 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659456 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659516 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659529 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659663 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.659701 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660057 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660089 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660099 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660326 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660344 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660431 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660409 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660501 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660508 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660631 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.660658 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.661130 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.661156 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.661164 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.699648 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.700694 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.700777 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.700842 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.700971 4492 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.701912 4492 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.25:6443: connect: connection refused" node="crc" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.715686 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="400ms" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736257 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736286 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736307 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736322 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736340 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736356 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736423 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736450 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736485 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736504 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736536 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736571 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736935 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.736993 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.737030 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.837954 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.837985 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838002 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838018 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838033 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838050 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838065 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838086 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838102 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838117 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838131 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838145 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838160 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838174 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838188 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838416 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838466 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838512 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838532 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838552 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838571 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838592 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838612 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838635 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838655 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838675 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838695 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838715 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838735 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.838756 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.903004 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.903905 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.903936 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.903949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.903971 4492 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 06:40:43 crc kubenswrapper[4492]: E0220 06:40:43.904257 4492 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.25:6443: connect: connection refused" node="crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.977770 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 20 06:40:43 crc kubenswrapper[4492]: I0220 06:40:43.983586 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 20 06:40:44 crc kubenswrapper[4492]: W0220 06:40:44.002701 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-87490a159411be01027b9737e014c692247c74c605f3e5bde84f5a397312ce19 WatchSource:0}: Error finding container 87490a159411be01027b9737e014c692247c74c605f3e5bde84f5a397312ce19: Status 404 returned error can't find the container with id 87490a159411be01027b9737e014c692247c74c605f3e5bde84f5a397312ce19 Feb 20 06:40:44 crc kubenswrapper[4492]: W0220 06:40:44.006291 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-504080df552251d10c2368748df3a84fa7d725eb9a6712101bc0d2a2e87a56bd WatchSource:0}: Error finding container 504080df552251d10c2368748df3a84fa7d725eb9a6712101bc0d2a2e87a56bd: Status 404 returned error can't find the container with id 504080df552251d10c2368748df3a84fa7d725eb9a6712101bc0d2a2e87a56bd Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.007899 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:44 crc kubenswrapper[4492]: W0220 06:40:44.016631 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-b2096119fb75d3cf32164bcd2038a519fc92d02f036143ef56742e7883e805c8 WatchSource:0}: Error finding container b2096119fb75d3cf32164bcd2038a519fc92d02f036143ef56742e7883e805c8: Status 404 returned error can't find the container with id b2096119fb75d3cf32164bcd2038a519fc92d02f036143ef56742e7883e805c8 Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.020639 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.025975 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:44 crc kubenswrapper[4492]: W0220 06:40:44.032565 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e786895424764cf0cdc777ef784bce8e1edc18501310a9d8386767d207f4d35b WatchSource:0}: Error finding container e786895424764cf0cdc777ef784bce8e1edc18501310a9d8386767d207f4d35b: Status 404 returned error can't find the container with id e786895424764cf0cdc777ef784bce8e1edc18501310a9d8386767d207f4d35b Feb 20 06:40:44 crc kubenswrapper[4492]: W0220 06:40:44.036962 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-de1be799a547b6efd00a4f1f04a4f312b7b66ba4231133e609cbcf227c0c4924 WatchSource:0}: Error finding container de1be799a547b6efd00a4f1f04a4f312b7b66ba4231133e609cbcf227c0c4924: Status 404 returned error can't find the container with id de1be799a547b6efd00a4f1f04a4f312b7b66ba4231133e609cbcf227c0c4924 Feb 20 06:40:44 crc kubenswrapper[4492]: E0220 06:40:44.116871 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="800ms" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.305342 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.306269 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.306301 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.306311 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.306341 4492 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 06:40:44 crc kubenswrapper[4492]: E0220 06:40:44.306750 4492 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.25:6443: connect: connection refused" node="crc" Feb 20 06:40:44 crc kubenswrapper[4492]: W0220 06:40:44.367160 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:44 crc kubenswrapper[4492]: E0220 06:40:44.367450 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.509808 4492 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.515093 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 07:17:25.513412179 +0000 UTC Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.561785 4492 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd" exitCode=0 Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.561866 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.561968 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"de1be799a547b6efd00a4f1f04a4f312b7b66ba4231133e609cbcf227c0c4924"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.562085 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.562961 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.563001 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e786895424764cf0cdc777ef784bce8e1edc18501310a9d8386767d207f4d35b"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.563067 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.563101 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.563111 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.564069 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534" exitCode=0 Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.564132 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.564150 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b2096119fb75d3cf32164bcd2038a519fc92d02f036143ef56742e7883e805c8"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.564222 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.564804 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.564834 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.564844 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.565942 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.566622 4492 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5682c4869082fa1acab86f75e69fd5d1ef91aa65f3f95ac466cec50693a79abf" exitCode=0 Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.566696 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5682c4869082fa1acab86f75e69fd5d1ef91aa65f3f95ac466cec50693a79abf"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.566715 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"87490a159411be01027b9737e014c692247c74c605f3e5bde84f5a397312ce19"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.566792 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.567172 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.567198 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.567210 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.567542 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.567571 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.567581 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.568314 4492 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="eda42d8dd09c47cbf9ee70cfa621c6eb7d03ec9277d91257c14322e73a0d5d1f" exitCode=0 Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.568342 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"eda42d8dd09c47cbf9ee70cfa621c6eb7d03ec9277d91257c14322e73a0d5d1f"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.568357 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"504080df552251d10c2368748df3a84fa7d725eb9a6712101bc0d2a2e87a56bd"} Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.568446 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.571771 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.571812 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:44 crc kubenswrapper[4492]: I0220 06:40:44.571826 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:44 crc kubenswrapper[4492]: E0220 06:40:44.917497 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="1.6s" Feb 20 06:40:45 crc kubenswrapper[4492]: W0220 06:40:45.030463 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:45 crc kubenswrapper[4492]: W0220 06:40:45.030689 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:45 crc kubenswrapper[4492]: E0220 06:40:45.030755 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:45 crc kubenswrapper[4492]: E0220 06:40:45.030751 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:45 crc kubenswrapper[4492]: W0220 06:40:45.071417 4492 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:40:45 crc kubenswrapper[4492]: E0220 06:40:45.071496 4492 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.25:6443: connect: connection refused" logger="UnhandledError" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.107029 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.107802 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.107826 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.107836 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.107856 4492 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 06:40:45 crc kubenswrapper[4492]: E0220 06:40:45.108116 4492 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.25:6443: connect: connection refused" node="crc" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.515213 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 23:32:28.186753698 +0000 UTC Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.549515 4492 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.572516 4492 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="08368f8dfa1ff094e57f746f499d5aa6f0fd9a2e1e94ef7339cb5f47d99bd96e" exitCode=0 Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.572523 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"08368f8dfa1ff094e57f746f499d5aa6f0fd9a2e1e94ef7339cb5f47d99bd96e"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.572708 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.573705 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.573733 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.573742 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.579290 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.579374 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.579393 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.579578 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.580828 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.580869 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.580881 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.587548 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.587599 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.587612 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.587705 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.588703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.588805 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.588889 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.590843 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.590872 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.590883 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.590891 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.590899 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.590980 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.591681 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.591709 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.591720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.593321 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a201f8dd6fd229ffc0d770b6340de35a4e97601ce09a8f5d3ac92bb67c574aa6"} Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.593390 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.594011 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.594029 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:45 crc kubenswrapper[4492]: I0220 06:40:45.594038 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.464404 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.515349 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 09:10:58.67336603 +0000 UTC Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.598640 4492 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4af8fd874cb5c56d88a7855415b20d5ece4d7ab723f6d779019be494a6c2451c" exitCode=0 Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.598696 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4af8fd874cb5c56d88a7855415b20d5ece4d7ab723f6d779019be494a6c2451c"} Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.598810 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.598852 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.598867 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.599897 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.599925 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.599943 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.599950 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.599959 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.599963 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.600171 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.600206 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.600223 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.709013 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.710069 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.710100 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.710112 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:46 crc kubenswrapper[4492]: I0220 06:40:46.710137 4492 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.515511 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 09:42:41.661181007 +0000 UTC Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.565133 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.565285 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.566373 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.566407 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.566417 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.606365 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"693404fcaf966a707f4d5ee5f72466873e33188f019eb07a819e52d774b489f8"} Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.606439 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c82f1f4abdf61156fe0997903135d59828c1913e7d5dc56ccd863c0af9854766"} Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.606452 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fa8901bcd48ad3e0f9ea540e2a05e1177ec64c608504a8f91e02f2a38950d44e"} Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.606466 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1ad97b83da8467370c50107a9d345df8e4f74b9d0b4396d3ee8899696be74272"} Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.606490 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f3a46e82a3a9c08ce424bb59b1720fe552bb5bddc3ae515a52eab7b71cfd9b49"} Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.606587 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.607176 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.607207 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.607215 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:47 crc kubenswrapper[4492]: I0220 06:40:47.938651 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.473730 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.473964 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.475085 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.475115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.475128 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.515948 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 11:28:22.80973529 +0000 UTC Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.608370 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.609456 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.609510 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:48 crc kubenswrapper[4492]: I0220 06:40:48.609523 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:49 crc kubenswrapper[4492]: I0220 06:40:49.516431 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 01:15:49.992120959 +0000 UTC Feb 20 06:40:49 crc kubenswrapper[4492]: I0220 06:40:49.604409 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 20 06:40:49 crc kubenswrapper[4492]: I0220 06:40:49.610604 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:49 crc kubenswrapper[4492]: I0220 06:40:49.611467 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:49 crc kubenswrapper[4492]: I0220 06:40:49.611522 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:49 crc kubenswrapper[4492]: I0220 06:40:49.611532 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.517441 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 16:19:01.363000845 +0000 UTC Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.525649 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.525791 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.526814 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.526861 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.526871 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.612744 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.613411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.613440 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:50 crc kubenswrapper[4492]: I0220 06:40:50.613448 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:51 crc kubenswrapper[4492]: I0220 06:40:51.517881 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 18:57:14.064758244 +0000 UTC Feb 20 06:40:52 crc kubenswrapper[4492]: I0220 06:40:52.517969 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 21:52:08.085885728 +0000 UTC Feb 20 06:40:52 crc kubenswrapper[4492]: I0220 06:40:52.965196 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:52 crc kubenswrapper[4492]: I0220 06:40:52.965341 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:52 crc kubenswrapper[4492]: I0220 06:40:52.966173 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:52 crc kubenswrapper[4492]: I0220 06:40:52.966201 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:52 crc kubenswrapper[4492]: I0220 06:40:52.966210 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:53 crc kubenswrapper[4492]: I0220 06:40:53.518865 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 05:18:19.331475607 +0000 UTC Feb 20 06:40:53 crc kubenswrapper[4492]: E0220 06:40:53.605136 4492 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.090727 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.090881 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.091901 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.091934 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.091943 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.256906 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.519385 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 11:15:52.638267314 +0000 UTC Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.620022 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.620705 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.620738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.620746 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.739520 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:54 crc kubenswrapper[4492]: I0220 06:40:54.743670 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.344397 4492 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.344457 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.348638 4492 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.348691 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.520051 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 00:58:06.362147615 +0000 UTC Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.621768 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.622730 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.622766 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.622776 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:55 crc kubenswrapper[4492]: I0220 06:40:55.625122 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:40:56 crc kubenswrapper[4492]: I0220 06:40:56.465792 4492 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 20 06:40:56 crc kubenswrapper[4492]: I0220 06:40:56.465863 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 20 06:40:56 crc kubenswrapper[4492]: I0220 06:40:56.520780 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 04:03:48.539495956 +0000 UTC Feb 20 06:40:56 crc kubenswrapper[4492]: I0220 06:40:56.623223 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:56 crc kubenswrapper[4492]: I0220 06:40:56.624087 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:56 crc kubenswrapper[4492]: I0220 06:40:56.624116 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:56 crc kubenswrapper[4492]: I0220 06:40:56.624125 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.090786 4492 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.090836 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.521786 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 22:22:45.936278848 +0000 UTC Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.624932 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.626073 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.626104 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.626114 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.957440 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.957678 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.958374 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.958404 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.958413 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:57 crc kubenswrapper[4492]: I0220 06:40:57.966425 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.479589 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.479742 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.479943 4492 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.479985 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.480577 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.480605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.480613 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.482792 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.522660 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 23:03:41.119586255 +0000 UTC Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.627196 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.627658 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.627930 4492 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.627994 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.628656 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.628693 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.628703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.628774 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.628805 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:40:58 crc kubenswrapper[4492]: I0220 06:40:58.628815 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:40:59 crc kubenswrapper[4492]: I0220 06:40:59.522796 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 07:49:32.136184914 +0000 UTC Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.328354 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.330924 4492 trace.go:236] Trace[2063245008]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 06:40:47.190) (total time: 13140ms): Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[2063245008]: ---"Objects listed" error: 13140ms (06:41:00.330) Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[2063245008]: [13.140335406s] [13.140335406s] END Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.330955 4492 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.331700 4492 trace.go:236] Trace[846431647]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 06:40:47.194) (total time: 13137ms): Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[846431647]: ---"Objects listed" error: 13137ms (06:41:00.331) Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[846431647]: [13.13758743s] [13.13758743s] END Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.331718 4492 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.332536 4492 trace.go:236] Trace[2010106556]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 06:40:46.929) (total time: 13403ms): Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[2010106556]: ---"Objects listed" error: 13403ms (06:41:00.332) Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[2010106556]: [13.403134234s] [13.403134234s] END Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.332574 4492 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.332902 4492 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.333594 4492 trace.go:236] Trace[2355853]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (20-Feb-2026 06:40:48.044) (total time: 12289ms): Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[2355853]: ---"Objects listed" error: 12288ms (06:41:00.333) Feb 20 06:41:00 crc kubenswrapper[4492]: Trace[2355853]: [12.289061897s] [12.289061897s] END Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.333614 4492 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.334616 4492 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.350559 4492 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.367969 4492 csr.go:261] certificate signing request csr-m9j57 is approved, waiting to be issued Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.373799 4492 csr.go:257] certificate signing request csr-m9j57 is issued Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.511836 4492 apiserver.go:52] "Watching apiserver" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.516129 4492 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.516510 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.516974 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.517095 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.517162 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.517240 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.517254 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.517360 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.517397 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.517632 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.517711 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.519152 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.520293 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.520323 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.520593 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.521977 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.522010 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.522819 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.522945 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 11:05:20.990533576 +0000 UTC Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.524650 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.531781 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.548928 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.561737 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.571181 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.584423 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.594209 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.606603 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.616353 4492 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.622930 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.635021 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636176 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636228 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636250 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636264 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636282 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636308 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636327 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636347 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636367 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636404 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636420 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636441 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636458 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636494 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636518 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636533 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636548 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636567 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636583 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636601 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636618 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636633 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636649 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636665 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636680 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636694 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636693 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636715 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636710 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636785 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636817 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636839 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636858 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636877 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636897 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636914 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636935 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636952 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636964 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636970 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.636970 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637015 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637035 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637053 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637074 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637095 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637112 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637130 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637143 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637152 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637167 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637184 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637189 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637203 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637228 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637246 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637261 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637283 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637302 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637319 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637337 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637355 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637366 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637373 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637462 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637513 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637540 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637568 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637592 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637614 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637639 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637662 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637686 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637712 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637737 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637757 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637784 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637805 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637828 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637850 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637870 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637897 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637932 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637953 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637984 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638010 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638034 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638052 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638070 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638088 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638108 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638134 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638154 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638172 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638190 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638212 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638231 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638248 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638267 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638283 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638305 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638325 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638343 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638369 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638386 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638409 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638429 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638450 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638468 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638505 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638525 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638544 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638564 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638583 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638601 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638621 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638642 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638658 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638677 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638696 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638712 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638733 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638750 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638768 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638786 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638808 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638827 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638848 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638868 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638888 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638908 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638930 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638954 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638974 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638993 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639013 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639038 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639056 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639074 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639095 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639113 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639136 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639167 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639188 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639206 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639232 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639250 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639269 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639289 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639333 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639415 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639437 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639454 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640187 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640214 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640232 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640250 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640271 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640293 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640329 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640352 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640372 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637590 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637724 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.637878 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638013 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638351 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638531 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638667 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.638831 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639133 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.639738 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640453 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640791 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640802 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.640993 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641119 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641397 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641403 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641497 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641526 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641605 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641659 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641688 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641830 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641966 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.641985 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.650877 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.642252 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.642286 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.642455 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.642531 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.642557 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.642698 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.642818 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.643063 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.643383 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.643423 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.643846 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.644318 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.644640 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.647571 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.647694 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.647859 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.647873 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.648096 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.648352 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.648379 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.648672 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.648896 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.648960 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.649306 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.649601 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.649796 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.649973 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.650164 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.650350 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.650548 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.650699 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.650840 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651090 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651123 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651361 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651548 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651820 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651851 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651872 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651895 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651918 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651939 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651959 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651963 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.651981 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652011 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652032 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652057 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652080 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652104 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652438 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652688 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.652926 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.653269 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.653611 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.653973 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.654263 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.654631 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.655946 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.656055 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.656518 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.656608 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.656738 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.656855 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.656869 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.656977 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.657183 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.657340 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.657525 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.657666 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.657806 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.658052 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.658272 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.658530 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.658656 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.658780 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.659115 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.659254 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.659395 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.659678 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.660069 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.661548 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.661779 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.661940 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.662318 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.662673 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.662813 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.662958 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.663746 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.664659 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.665417 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.665748 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.665981 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.666083 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.666351 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.666761 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.667738 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:41:01.167717503 +0000 UTC m=+17.939006481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.667624 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.670960 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.671067 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.671177 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.671269 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.671671 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.671796 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.671889 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.671977 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672066 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672150 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672234 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672318 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672420 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672547 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672645 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672730 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672822 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.672917 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.673009 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.673129 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.673220 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.673312 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.673420 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674079 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674184 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674273 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674437 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674585 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674736 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674847 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.674965 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675062 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675226 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675320 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675419 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675547 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675669 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675771 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675868 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.675958 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676221 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676306 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676407 4492 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676529 4492 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676609 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676814 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676911 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.676989 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677070 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677128 4492 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677207 4492 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677280 4492 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677357 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677446 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677539 4492 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677615 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677685 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677741 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677819 4492 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677894 4492 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.677969 4492 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678021 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678088 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678174 4492 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678256 4492 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678342 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678422 4492 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678508 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678623 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678703 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678774 4492 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678860 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678937 4492 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.678994 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679068 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679144 4492 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679226 4492 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679277 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679399 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679496 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679578 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679713 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679793 4492 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.679921 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680004 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680121 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680203 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680254 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680322 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680403 4492 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680490 4492 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680569 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680646 4492 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680697 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680815 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680895 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.680947 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681418 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681524 4492 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681613 4492 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681694 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681763 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681825 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681892 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.681956 4492 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682021 4492 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682098 4492 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682172 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682238 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682300 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682349 4492 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682427 4492 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.682612 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.685530 4492 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.685642 4492 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.685731 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.685812 4492 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.685887 4492 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.685960 4492 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686014 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686096 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686175 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686255 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686329 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686410 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686993 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687095 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687182 4492 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687290 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687417 4492 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687603 4492 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.688107 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.688636 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.688722 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.688809 4492 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.688874 4492 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.688934 4492 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689004 4492 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689069 4492 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689148 4492 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689205 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689274 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689340 4492 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689414 4492 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689490 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689560 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689625 4492 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689684 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689748 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689816 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689877 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689935 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689988 4492 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.690060 4492 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.690124 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.683692 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.683782 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.684720 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.684885 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.686723 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687195 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687356 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687607 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.687850 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.688072 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689344 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689762 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.689980 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.690281 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.691282 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.695411 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.695782 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.695986 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.696110 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.696890 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.697122 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.697597 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.698831 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.699462 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.699623 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.700771 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.700944 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.702418 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.702597 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.702898 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.703256 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.703809 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.704018 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.704206 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.709229 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.709457 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.709530 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.709875 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.709907 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.710042 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.710176 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.710379 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.710522 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.710661 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.710854 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.710978 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711184 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711381 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711516 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711537 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711575 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711690 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711717 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711844 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.711854 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.712219 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.712407 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.712601 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.712793 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.713078 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.713252 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.713535 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.713789 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.713946 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.714068 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.714332 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.714511 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.714876 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.715079 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.715233 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.715415 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.715544 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.715613 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.715879 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.716095 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.717050 4492 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.720290 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.721129 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.729992 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.730303 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.730366 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:01.230350362 +0000 UTC m=+18.001639340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.730444 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.730490 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:01.230468827 +0000 UTC m=+18.001757806 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.730443 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.731498 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.735750 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.742871 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.743404 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.746132 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.746157 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.746169 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.746213 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:01.246200063 +0000 UTC m=+18.017489041 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.750926 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.751780 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.751803 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.751814 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:00 crc kubenswrapper[4492]: E0220 06:41:00.751871 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:01.251849562 +0000 UTC m=+18.023138540 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.756007 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.764831 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.790826 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.790966 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791025 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.790995 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791234 4492 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791265 4492 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791277 4492 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791287 4492 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791297 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791306 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791316 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791325 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791336 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791345 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791353 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791362 4492 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791370 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791380 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791388 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791404 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791413 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791424 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791433 4492 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791443 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791452 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791461 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791490 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791500 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791512 4492 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791521 4492 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791529 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791538 4492 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791546 4492 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791554 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791562 4492 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791569 4492 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791577 4492 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791585 4492 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791592 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791601 4492 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791610 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791618 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791626 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791634 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791642 4492 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791649 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791657 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791664 4492 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791674 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791683 4492 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791692 4492 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791701 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791709 4492 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791718 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791726 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791734 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791743 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791750 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791758 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791766 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791774 4492 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791783 4492 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791790 4492 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791797 4492 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791805 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791816 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791824 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791832 4492 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791839 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791848 4492 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791855 4492 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791862 4492 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791869 4492 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791878 4492 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791886 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791893 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791901 4492 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791908 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791917 4492 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791925 4492 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791932 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791939 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.791947 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.831817 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.837578 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 20 06:41:00 crc kubenswrapper[4492]: W0220 06:41:00.842643 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-bffd6e77ca06ed7261b83f22f098a9cb34a999885b0db3f598306be81eb85404 WatchSource:0}: Error finding container bffd6e77ca06ed7261b83f22f098a9cb34a999885b0db3f598306be81eb85404: Status 404 returned error can't find the container with id bffd6e77ca06ed7261b83f22f098a9cb34a999885b0db3f598306be81eb85404 Feb 20 06:41:00 crc kubenswrapper[4492]: I0220 06:41:00.843506 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 20 06:41:00 crc kubenswrapper[4492]: W0220 06:41:00.850992 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-61255bf913d1629371842ab2ae45f635eeca8fa6521dddf1085c378b800a8562 WatchSource:0}: Error finding container 61255bf913d1629371842ab2ae45f635eeca8fa6521dddf1085c378b800a8562: Status 404 returned error can't find the container with id 61255bf913d1629371842ab2ae45f635eeca8fa6521dddf1085c378b800a8562 Feb 20 06:41:00 crc kubenswrapper[4492]: W0220 06:41:00.856184 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-f7e2e739c2ed32762653043d108bdb6a5011e745b8bddc0ab356209b8ed14f04 WatchSource:0}: Error finding container f7e2e739c2ed32762653043d108bdb6a5011e745b8bddc0ab356209b8ed14f04: Status 404 returned error can't find the container with id f7e2e739c2ed32762653043d108bdb6a5011e745b8bddc0ab356209b8ed14f04 Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.028549 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-xbm4p"] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.028822 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.029910 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.030512 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.032427 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.041706 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.056993 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.087086 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.100104 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.120576 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.148227 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.173459 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.202932 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.203013 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbqjh\" (UniqueName: \"kubernetes.io/projected/8b630176-8730-4b5a-bf14-46bf2bc2862d-kube-api-access-cbqjh\") pod \"node-resolver-xbm4p\" (UID: \"8b630176-8730-4b5a-bf14-46bf2bc2862d\") " pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.203071 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:41:02.203050577 +0000 UTC m=+18.974339555 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.203119 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8b630176-8730-4b5a-bf14-46bf2bc2862d-hosts-file\") pod \"node-resolver-xbm4p\" (UID: \"8b630176-8730-4b5a-bf14-46bf2bc2862d\") " pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.304320 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbqjh\" (UniqueName: \"kubernetes.io/projected/8b630176-8730-4b5a-bf14-46bf2bc2862d-kube-api-access-cbqjh\") pod \"node-resolver-xbm4p\" (UID: \"8b630176-8730-4b5a-bf14-46bf2bc2862d\") " pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.304364 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.304388 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.304421 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8b630176-8730-4b5a-bf14-46bf2bc2862d-hosts-file\") pod \"node-resolver-xbm4p\" (UID: \"8b630176-8730-4b5a-bf14-46bf2bc2862d\") " pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.304442 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.304461 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304549 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304613 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:02.304597619 +0000 UTC m=+19.075886597 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304615 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304647 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304661 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.304666 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8b630176-8730-4b5a-bf14-46bf2bc2862d-hosts-file\") pod \"node-resolver-xbm4p\" (UID: \"8b630176-8730-4b5a-bf14-46bf2bc2862d\") " pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304703 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304617 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304735 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304754 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304714 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:02.304699844 +0000 UTC m=+19.075988812 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304828 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:02.304810283 +0000 UTC m=+19.076099271 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:01 crc kubenswrapper[4492]: E0220 06:41:01.304854 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:02.304847163 +0000 UTC m=+19.076136152 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.323946 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbqjh\" (UniqueName: \"kubernetes.io/projected/8b630176-8730-4b5a-bf14-46bf2bc2862d-kube-api-access-cbqjh\") pod \"node-resolver-xbm4p\" (UID: \"8b630176-8730-4b5a-bf14-46bf2bc2862d\") " pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.337285 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-xbm4p" Feb 20 06:41:01 crc kubenswrapper[4492]: W0220 06:41:01.346285 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b630176_8730_4b5a_bf14_46bf2bc2862d.slice/crio-294a3884f2f00c894aade0f610ac392629043f3b2758328f28bc5fffd5711929 WatchSource:0}: Error finding container 294a3884f2f00c894aade0f610ac392629043f3b2758328f28bc5fffd5711929: Status 404 returned error can't find the container with id 294a3884f2f00c894aade0f610ac392629043f3b2758328f28bc5fffd5711929 Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.375654 4492 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-20 06:36:00 +0000 UTC, rotation deadline is 2027-01-03 06:11:54.768431733 +0000 UTC Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.375704 4492 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7607h30m53.392729987s for next certificate rotation Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.504441 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-zrlsb"] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.504861 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.506933 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.507545 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.508220 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.510060 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.510182 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.516736 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.523327 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 11:07:22.831439993 +0000 UTC Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.534705 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.545512 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.553005 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.559567 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.560282 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.560976 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.561635 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.562257 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.562820 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.562795 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.564244 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.564784 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.565745 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.566201 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.567023 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.567619 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.568410 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.568870 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.569662 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.570129 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.570648 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.571315 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.571834 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.572335 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.572696 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.573080 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.573590 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.574289 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.574976 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.575379 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.576285 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.577200 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.577654 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.578129 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.578931 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.579353 4492 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.579459 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.581327 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.581799 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.581851 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.582183 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.583546 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.584637 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.585167 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.585861 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.586501 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.586927 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.587497 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.588082 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.589950 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.590290 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.590394 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.591235 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.591764 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.592965 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.593396 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.593830 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.594690 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.595151 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.596022 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.596450 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.606792 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f8755354-8581-4b5d-9b7e-6b7c53707f8c-proxy-tls\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.606839 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8755354-8581-4b5d-9b7e-6b7c53707f8c-mcd-auth-proxy-config\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.606863 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqc6l\" (UniqueName: \"kubernetes.io/projected/f8755354-8581-4b5d-9b7e-6b7c53707f8c-kube-api-access-nqc6l\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.606898 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f8755354-8581-4b5d-9b7e-6b7c53707f8c-rootfs\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.636634 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.638510 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26" exitCode=255 Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.638565 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.642458 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xbm4p" event={"ID":"8b630176-8730-4b5a-bf14-46bf2bc2862d","Type":"ContainerStarted","Data":"418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.642514 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-xbm4p" event={"ID":"8b630176-8730-4b5a-bf14-46bf2bc2862d","Type":"ContainerStarted","Data":"294a3884f2f00c894aade0f610ac392629043f3b2758328f28bc5fffd5711929"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.643427 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f7e2e739c2ed32762653043d108bdb6a5011e745b8bddc0ab356209b8ed14f04"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.645180 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.645238 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.645253 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"61255bf913d1629371842ab2ae45f635eeca8fa6521dddf1085c378b800a8562"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.646867 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.646930 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bffd6e77ca06ed7261b83f22f098a9cb34a999885b0db3f598306be81eb85404"} Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.650278 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.662596 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.672501 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.680332 4492 scope.go:117] "RemoveContainer" containerID="cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.680990 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.684691 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.704685 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.707437 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f8755354-8581-4b5d-9b7e-6b7c53707f8c-proxy-tls\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.707537 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8755354-8581-4b5d-9b7e-6b7c53707f8c-mcd-auth-proxy-config\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.707566 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqc6l\" (UniqueName: \"kubernetes.io/projected/f8755354-8581-4b5d-9b7e-6b7c53707f8c-kube-api-access-nqc6l\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.707601 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f8755354-8581-4b5d-9b7e-6b7c53707f8c-rootfs\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.707679 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f8755354-8581-4b5d-9b7e-6b7c53707f8c-rootfs\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.708343 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f8755354-8581-4b5d-9b7e-6b7c53707f8c-mcd-auth-proxy-config\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.711186 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f8755354-8581-4b5d-9b7e-6b7c53707f8c-proxy-tls\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.721141 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.726600 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqc6l\" (UniqueName: \"kubernetes.io/projected/f8755354-8581-4b5d-9b7e-6b7c53707f8c-kube-api-access-nqc6l\") pod \"machine-config-daemon-zrlsb\" (UID: \"f8755354-8581-4b5d-9b7e-6b7c53707f8c\") " pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.738732 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.765257 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.776860 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.790501 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.805174 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.821636 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.822806 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:41:01 crc kubenswrapper[4492]: W0220 06:41:01.831964 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8755354_8581_4b5d_9b7e_6b7c53707f8c.slice/crio-c02853e1d50167d7581808758428ce9b4bd359d4a07b62c5d162444f48b13935 WatchSource:0}: Error finding container c02853e1d50167d7581808758428ce9b4bd359d4a07b62c5d162444f48b13935: Status 404 returned error can't find the container with id c02853e1d50167d7581808758428ce9b4bd359d4a07b62c5d162444f48b13935 Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.840910 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.853417 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.867348 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.883733 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-pl2t8"] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.884200 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2rr9j"] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.884373 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2rr9j" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.884633 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.887670 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.888011 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f68mj"] Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.888664 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.892691 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.893288 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.893560 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.893657 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.893689 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.894798 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.894868 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.895024 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.895027 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.895039 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.895146 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.895209 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.895245 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.895646 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.921696 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.941297 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.953002 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.967703 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.982510 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:01 crc kubenswrapper[4492]: I0220 06:41:01.990275 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:01Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.002948 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009230 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-systemd\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009269 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-ovn\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009294 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-conf-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009315 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8khkx\" (UniqueName: \"kubernetes.io/projected/dcc11973-022a-47f9-b1e2-23e945352e7f-kube-api-access-8khkx\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009335 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-kubelet\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009360 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-etc-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009390 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-cnibin\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009418 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-kubelet\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009507 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-slash\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009546 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009565 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-ovn-kubernetes\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009599 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-netd\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009619 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-config\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009638 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhn54\" (UniqueName: \"kubernetes.io/projected/28d6c67f-c4cd-4692-b490-b2b884c72db6-kube-api-access-bhn54\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009659 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-cni-multus\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009676 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-node-log\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009699 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-script-lib\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009719 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-cni-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009734 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-multus-certs\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009769 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-k8s-cni-cncf-io\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009815 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-hostroot\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009854 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-etc-kubernetes\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009886 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-system-cni-dir\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009903 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-systemd-units\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009934 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-var-lib-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.009978 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-system-cni-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010003 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010027 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgw7t\" (UniqueName: \"kubernetes.io/projected/65a8a150-0190-4595-8538-2c4c5875ba54-kube-api-access-qgw7t\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010062 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-socket-dir-parent\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010084 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65a8a150-0190-4595-8538-2c4c5875ba54-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010100 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovn-node-metrics-cert\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010168 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-cnibin\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010216 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-netns\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010241 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-bin\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010264 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010285 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-env-overrides\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010331 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-os-release\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010352 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-cni-bin\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010401 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-os-release\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010434 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-netns\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010453 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-daemon-config\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010490 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-log-socket\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010518 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dcc11973-022a-47f9-b1e2-23e945352e7f-cni-binary-copy\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.010562 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65a8a150-0190-4595-8538-2c4c5875ba54-cni-binary-copy\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.014245 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.033773 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.067511 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.106641 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111014 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-systemd\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111049 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-ovn\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111073 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-conf-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111093 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8khkx\" (UniqueName: \"kubernetes.io/projected/dcc11973-022a-47f9-b1e2-23e945352e7f-kube-api-access-8khkx\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111112 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-kubelet\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111134 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-etc-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111151 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-slash\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111168 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111184 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-ovn-kubernetes\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111216 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-cnibin\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111233 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-kubelet\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111251 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-netd\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111269 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-config\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111294 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhn54\" (UniqueName: \"kubernetes.io/projected/28d6c67f-c4cd-4692-b490-b2b884c72db6-kube-api-access-bhn54\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111310 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-cni-multus\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111327 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-node-log\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111348 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-script-lib\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111366 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-multus-certs\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111385 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-cni-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111400 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-k8s-cni-cncf-io\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111425 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-hostroot\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111439 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-etc-kubernetes\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111459 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-system-cni-dir\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111489 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-systemd-units\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111510 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-var-lib-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111526 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111549 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-system-cni-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111578 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgw7t\" (UniqueName: \"kubernetes.io/projected/65a8a150-0190-4595-8538-2c4c5875ba54-kube-api-access-qgw7t\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111595 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65a8a150-0190-4595-8538-2c4c5875ba54-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111610 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovn-node-metrics-cert\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111633 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-socket-dir-parent\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111648 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-cnibin\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111662 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-netns\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111683 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-bin\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111702 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111716 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-env-overrides\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111731 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-netns\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111759 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-os-release\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111775 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-cni-bin\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111797 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-os-release\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111813 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-daemon-config\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111828 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-log-socket\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111844 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dcc11973-022a-47f9-b1e2-23e945352e7f-cni-binary-copy\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.111858 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65a8a150-0190-4595-8538-2c4c5875ba54-cni-binary-copy\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112183 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-systemd\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112259 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-ovn\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112297 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-conf-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112622 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65a8a150-0190-4595-8538-2c4c5875ba54-cni-binary-copy\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112643 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-kubelet\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112679 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-etc-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112691 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-var-lib-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112721 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-slash\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112745 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-openvswitch\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112768 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-ovn-kubernetes\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112788 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-cnibin\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112810 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-kubelet\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112831 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-netd\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.112941 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-system-cni-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113067 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113244 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-netns\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113347 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-config\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113457 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-cni-bin\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113550 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-env-overrides\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113559 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-os-release\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113573 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-os-release\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113588 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-var-lib-cni-multus\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113627 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-etc-kubernetes\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113660 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-k8s-cni-cncf-io\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113687 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-hostroot\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113722 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113806 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-cni-dir\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113862 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65a8a150-0190-4595-8538-2c4c5875ba54-system-cni-dir\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113888 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-systemd-units\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113903 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-cnibin\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113938 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-socket-dir-parent\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113950 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-netns\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.113977 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-node-log\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.114057 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65a8a150-0190-4595-8538-2c4c5875ba54-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.114113 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-log-socket\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.114139 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dcc11973-022a-47f9-b1e2-23e945352e7f-host-run-multus-certs\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.114157 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-bin\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.114244 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dcc11973-022a-47f9-b1e2-23e945352e7f-cni-binary-copy\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.114405 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dcc11973-022a-47f9-b1e2-23e945352e7f-multus-daemon-config\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.114556 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-script-lib\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.119061 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovn-node-metrics-cert\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.127528 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhn54\" (UniqueName: \"kubernetes.io/projected/28d6c67f-c4cd-4692-b490-b2b884c72db6-kube-api-access-bhn54\") pod \"ovnkube-node-f68mj\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.136940 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8khkx\" (UniqueName: \"kubernetes.io/projected/dcc11973-022a-47f9-b1e2-23e945352e7f-kube-api-access-8khkx\") pod \"multus-2rr9j\" (UID: \"dcc11973-022a-47f9-b1e2-23e945352e7f\") " pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.137975 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.140755 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgw7t\" (UniqueName: \"kubernetes.io/projected/65a8a150-0190-4595-8538-2c4c5875ba54-kube-api-access-qgw7t\") pod \"multus-additional-cni-plugins-pl2t8\" (UID: \"65a8a150-0190-4595-8538-2c4c5875ba54\") " pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.155549 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.198782 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2rr9j" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.204788 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.209359 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.212995 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.213319 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:41:04.213271089 +0000 UTC m=+20.984560057 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:41:02 crc kubenswrapper[4492]: W0220 06:41:02.216697 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65a8a150_0190_4595_8538_2c4c5875ba54.slice/crio-0f6cb203f2bb4304dd81c8c3e723d2391bba58ca1feb6917c898ae39810e446b WatchSource:0}: Error finding container 0f6cb203f2bb4304dd81c8c3e723d2391bba58ca1feb6917c898ae39810e446b: Status 404 returned error can't find the container with id 0f6cb203f2bb4304dd81c8c3e723d2391bba58ca1feb6917c898ae39810e446b Feb 20 06:41:02 crc kubenswrapper[4492]: W0220 06:41:02.217190 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcc11973_022a_47f9_b1e2_23e945352e7f.slice/crio-f8e32f702496376483eedd1e26685ce4ab1d22615151aa991f930082323433f1 WatchSource:0}: Error finding container f8e32f702496376483eedd1e26685ce4ab1d22615151aa991f930082323433f1: Status 404 returned error can't find the container with id f8e32f702496376483eedd1e26685ce4ab1d22615151aa991f930082323433f1 Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.313660 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.313757 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.313824 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.313861 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313865 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313891 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313905 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313916 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313942 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313957 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313953 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:04.313937299 +0000 UTC m=+21.085226277 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.313978 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.314004 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:04.31399086 +0000 UTC m=+21.085279839 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.314016 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.314022 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:04.314015057 +0000 UTC m=+21.085304035 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.314043 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:04.314036397 +0000 UTC m=+21.085325375 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.524436 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 21:46:52.677972894 +0000 UTC Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.555863 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.555884 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.555963 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.555981 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.556070 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:02 crc kubenswrapper[4492]: E0220 06:41:02.556131 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.651496 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87" exitCode=0 Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.651551 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.651577 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"1b30d5ac2f119de9934e653a72f38b71d982caa643ae28dd8acede7d1b09843b"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.653975 4492 generic.go:334] "Generic (PLEG): container finished" podID="65a8a150-0190-4595-8538-2c4c5875ba54" containerID="c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512" exitCode=0 Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.654040 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerDied","Data":"c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.654072 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerStarted","Data":"0f6cb203f2bb4304dd81c8c3e723d2391bba58ca1feb6917c898ae39810e446b"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.656831 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.658746 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.658873 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.660843 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.666539 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerStarted","Data":"d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.666569 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerStarted","Data":"f8e32f702496376483eedd1e26685ce4ab1d22615151aa991f930082323433f1"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.669223 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.669394 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.669708 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.669725 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"c02853e1d50167d7581808758428ce9b4bd359d4a07b62c5d162444f48b13935"} Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.687913 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.701623 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.716630 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.735507 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.745717 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.768704 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.781707 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.802978 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.818022 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.829922 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.838825 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.852168 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.864438 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.882748 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.901009 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.927269 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.942048 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.961458 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:02 crc kubenswrapper[4492]: I0220 06:41:02.987867 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:02Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.010227 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.020261 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.032213 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.052018 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.440425 4492 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.524709 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 10:32:45.97128885 +0000 UTC Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.533309 4492 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.534962 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.534997 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.535007 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.535091 4492 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.540204 4492 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.540359 4492 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.541115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.541153 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.541164 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.541179 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.541189 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: E0220 06:41:03.553964 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.557394 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.557543 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.557605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.557688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.557762 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.565911 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: E0220 06:41:03.565979 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.568247 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.568279 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.568290 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.568303 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.568315 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.576217 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: E0220 06:41:03.578018 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.580878 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.580929 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.580940 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.580957 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.580969 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.588993 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: E0220 06:41:03.590288 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.593453 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.593501 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.593514 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.593530 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.593542 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.598677 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: E0220 06:41:03.603882 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: E0220 06:41:03.603995 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.609176 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.611329 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.611345 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.611380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.611392 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.616028 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.626231 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.636022 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.645713 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.654287 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.664249 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.672245 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.675054 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.675090 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.675102 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.675112 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.675121 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.675133 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.677504 4492 generic.go:334] "Generic (PLEG): container finished" podID="65a8a150-0190-4595-8538-2c4c5875ba54" containerID="3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263" exitCode=0 Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.677622 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerDied","Data":"3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.685760 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.696336 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.705125 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.713232 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.713268 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.713281 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.713297 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.713312 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.720401 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.732168 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.746326 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.751863 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-p8vnd"] Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.753401 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.755094 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.755281 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.755719 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.756261 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.756327 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.767108 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.778034 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.792027 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.802370 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.812877 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.817202 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.817246 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.817259 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.817281 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.817295 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.824323 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.843221 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.854890 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.864589 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.874331 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.884267 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.895164 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.904635 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.913501 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.919660 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.919692 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.919705 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.919722 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.919735 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:03Z","lastTransitionTime":"2026-02-20T06:41:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.923404 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.928777 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2862f4a6-65d4-4621-acb8-78b9c13906dc-host\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.928824 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gtfn\" (UniqueName: \"kubernetes.io/projected/2862f4a6-65d4-4621-acb8-78b9c13906dc-kube-api-access-2gtfn\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.928867 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2862f4a6-65d4-4621-acb8-78b9c13906dc-serviceca\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.931498 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.955357 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:03 crc kubenswrapper[4492]: I0220 06:41:03.994073 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.021870 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.021899 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.021911 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.021927 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.021938 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.030180 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2862f4a6-65d4-4621-acb8-78b9c13906dc-serviceca\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.030257 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2862f4a6-65d4-4621-acb8-78b9c13906dc-host\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.030289 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gtfn\" (UniqueName: \"kubernetes.io/projected/2862f4a6-65d4-4621-acb8-78b9c13906dc-kube-api-access-2gtfn\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.030348 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2862f4a6-65d4-4621-acb8-78b9c13906dc-host\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.031060 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2862f4a6-65d4-4621-acb8-78b9c13906dc-serviceca\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.032580 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.060883 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gtfn\" (UniqueName: \"kubernetes.io/projected/2862f4a6-65d4-4621-acb8-78b9c13906dc-kube-api-access-2gtfn\") pod \"node-ca-p8vnd\" (UID: \"2862f4a6-65d4-4621-acb8-78b9c13906dc\") " pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.067671 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-p8vnd" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.094208 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.098390 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.102420 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.119271 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.123818 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.123856 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.123866 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.123879 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.123891 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.153467 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.194463 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.226343 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.226376 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.226386 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.226401 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.226412 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.231754 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.231955 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:41:08.231931082 +0000 UTC m=+25.003220060 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.232340 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.272272 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.313032 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.328770 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.328801 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.328809 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.328824 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.328836 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.333180 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.333220 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.333241 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.333270 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333351 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333372 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333389 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:08.333377824 +0000 UTC m=+25.104666802 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333395 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333408 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333458 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:08.333435594 +0000 UTC m=+25.104724572 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333516 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333539 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:08.333531967 +0000 UTC m=+25.104820935 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333583 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333593 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333600 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.333625 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:08.333613131 +0000 UTC m=+25.104902109 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.354892 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.392328 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.431494 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.431526 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.431537 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.431553 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.431563 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.435424 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.480857 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.524846 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 16:52:17.339456862 +0000 UTC Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.533829 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.533871 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.533881 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.533898 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.533909 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.550748 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.556436 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.556436 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.556581 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.556662 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.556438 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:04 crc kubenswrapper[4492]: E0220 06:41:04.556735 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.568363 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.592807 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.632090 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.635557 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.635589 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.635599 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.635615 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.635625 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.673736 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.682037 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-p8vnd" event={"ID":"2862f4a6-65d4-4621-acb8-78b9c13906dc","Type":"ContainerStarted","Data":"3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.682114 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-p8vnd" event={"ID":"2862f4a6-65d4-4621-acb8-78b9c13906dc","Type":"ContainerStarted","Data":"d0db79eb03b14f1b815f8c322c22b53faab3fe60d47693ff1dd6368132ea1d05"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.684530 4492 generic.go:334] "Generic (PLEG): container finished" podID="65a8a150-0190-4595-8538-2c4c5875ba54" containerID="83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0" exitCode=0 Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.684629 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerDied","Data":"83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.715487 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.738571 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.738608 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.738619 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.738635 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.738646 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.755103 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.801044 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.837193 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.840975 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.841025 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.841037 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.841054 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.841068 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.876752 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.914600 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.943733 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.943778 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.943790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.943810 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.943823 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:04Z","lastTransitionTime":"2026-02-20T06:41:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.954320 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:04 crc kubenswrapper[4492]: I0220 06:41:04.994649 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:04Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.036090 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.045517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.045546 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.045557 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.045570 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.045580 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.072292 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.111969 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.147745 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.147780 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.147790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.147807 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.147817 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.154164 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.194415 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.239039 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.249285 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.249316 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.249324 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.249337 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.249345 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.274076 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.314659 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.351254 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.351285 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.351295 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.351312 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.351321 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.356886 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.392779 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.434784 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.453704 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.453738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.453747 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.453763 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.453772 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.471587 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.512525 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.525746 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 04:49:42.552654009 +0000 UTC Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.555686 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.556244 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.556297 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.556329 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.556356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.556372 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.593758 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.632254 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.659649 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.659691 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.659704 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.659726 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.659740 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.675528 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.692440 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.694578 4492 generic.go:334] "Generic (PLEG): container finished" podID="65a8a150-0190-4595-8538-2c4c5875ba54" containerID="f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422" exitCode=0 Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.694623 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerDied","Data":"f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.717793 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.754835 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.761591 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.761635 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.761645 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.761662 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.761675 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.793607 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.832636 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.864821 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.864887 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.864905 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.864930 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.864948 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.879640 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.914522 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.956733 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.967739 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.967772 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.967782 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.967802 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:05 crc kubenswrapper[4492]: I0220 06:41:05.967814 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:05Z","lastTransitionTime":"2026-02-20T06:41:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.000226 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.034095 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.070191 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.070218 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.070228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.070244 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.070258 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.072504 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.113016 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.151768 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.172423 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.172460 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.172497 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.172514 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.172527 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.193578 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.233340 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.272418 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.275174 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.275202 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.275212 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.275226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.275236 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.377320 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.377362 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.377371 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.377384 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.377393 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.484860 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.484927 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.484950 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.484973 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.484985 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.526110 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 10:31:49.012307796 +0000 UTC Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.556549 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:06 crc kubenswrapper[4492]: E0220 06:41:06.556662 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.557414 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:06 crc kubenswrapper[4492]: E0220 06:41:06.557513 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.557584 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:06 crc kubenswrapper[4492]: E0220 06:41:06.557751 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.587408 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.587435 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.587444 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.587458 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.587469 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.689720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.689747 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.689757 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.689772 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.689784 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.700373 4492 generic.go:334] "Generic (PLEG): container finished" podID="65a8a150-0190-4595-8538-2c4c5875ba54" containerID="dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691" exitCode=0 Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.700420 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerDied","Data":"dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.711738 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.734612 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.745343 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.756611 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.768452 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.778585 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.789529 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.792176 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.792216 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.792227 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.792245 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.792258 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.799170 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.809388 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.818317 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.827635 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.838274 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.848674 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.863680 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:06Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.894772 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.894877 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.894940 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.895008 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.895066 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.997281 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.997319 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.997329 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.997349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:06 crc kubenswrapper[4492]: I0220 06:41:06.997358 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:06Z","lastTransitionTime":"2026-02-20T06:41:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.100068 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.100109 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.100119 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.100136 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.100149 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.202572 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.202620 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.202633 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.202656 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.202670 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.305484 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.305541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.305551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.305572 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.305584 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.408045 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.408084 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.408094 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.408110 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.408121 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.510098 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.510139 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.510149 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.510165 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.510175 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.526368 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 15:52:45.457263946 +0000 UTC Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.612453 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.612532 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.612544 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.612565 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.612578 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.712032 4492 generic.go:334] "Generic (PLEG): container finished" podID="65a8a150-0190-4595-8538-2c4c5875ba54" containerID="a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4" exitCode=0 Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.712178 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerDied","Data":"a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.714228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.714256 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.714265 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.714281 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.714295 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.732433 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.745242 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.755557 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.768544 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.782223 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.795989 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.805054 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.813736 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.816670 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.816708 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.816719 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.816739 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.816753 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.822795 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.834706 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.843922 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.855930 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.868581 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.879655 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:07Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.919043 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.919078 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.919090 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.919107 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:07 crc kubenswrapper[4492]: I0220 06:41:07.919119 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:07Z","lastTransitionTime":"2026-02-20T06:41:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.022003 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.022037 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.022048 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.022061 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.022070 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.123969 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.124011 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.124024 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.124039 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.124050 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.225818 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.225860 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.225868 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.225885 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.225896 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.273206 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.273433 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:41:16.273405987 +0000 UTC m=+33.044694975 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.327801 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.327839 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.327849 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.327866 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.327877 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.374188 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.374231 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.374251 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.374272 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374377 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374394 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374426 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:16.374414276 +0000 UTC m=+33.145703254 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374430 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374454 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:16.374438973 +0000 UTC m=+33.145727952 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374456 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374518 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374397 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374569 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374585 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374570 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:16.374554473 +0000 UTC m=+33.145843461 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.374683 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:16.374645596 +0000 UTC m=+33.145934574 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.429871 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.430201 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.430212 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.430228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.430242 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.526979 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 19:36:56.430008723 +0000 UTC Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.532577 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.532609 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.532619 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.532640 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.532652 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.556302 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.556317 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.556452 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.556689 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.556921 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:08 crc kubenswrapper[4492]: E0220 06:41:08.557040 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.634443 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.634661 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.634724 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.634782 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.634839 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.721071 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.721496 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.721529 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.725292 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" event={"ID":"65a8a150-0190-4595-8538-2c4c5875ba54","Type":"ContainerStarted","Data":"111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.732062 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.739308 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.739337 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.739346 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.739358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.739367 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.741213 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.750212 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.751420 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.757674 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.766321 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.773818 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.782989 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.789406 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.800733 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.810365 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.819482 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.828468 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.837867 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.841547 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.841573 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.841584 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.841598 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.841609 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.848680 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.861437 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.870290 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.879738 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.888792 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.901594 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.910191 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.919350 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.929866 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.944209 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.944249 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.944261 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.944282 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.944294 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:08Z","lastTransitionTime":"2026-02-20T06:41:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.944847 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.968349 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:08 crc kubenswrapper[4492]: I0220 06:41:08.982540 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:08Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.003260 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.019788 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.029230 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.037366 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.046779 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.046808 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.046817 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.046830 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.046840 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.149427 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.149464 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.149494 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.149508 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.149530 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.252057 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.252086 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.252095 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.252111 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.252120 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.354396 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.354432 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.354443 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.354462 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.354495 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.456529 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.456576 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.456587 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.456605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.456618 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.527289 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 15:51:42.486303248 +0000 UTC Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.559026 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.559049 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.559058 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.559070 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.559079 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.661375 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.661408 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.661419 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.661433 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.661443 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.734313 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/0.log" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.737581 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa" exitCode=1 Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.737621 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.738155 4492 scope.go:117] "RemoveContainer" containerID="8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.749636 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.761098 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.763743 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.763778 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.763788 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.763802 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.763814 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.772008 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.783950 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.800243 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:09Z\\\",\\\"message\\\":\\\"220 06:41:09.672966 5704 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:09.673130 5704 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:09.673395 5704 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 06:41:09.673660 5704 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:09.673837 5704 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 06:41:09.673945 5704 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 06:41:09.673997 5704 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 06:41:09.674007 5704 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 06:41:09.674042 5704 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 06:41:09.674054 5704 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 06:41:09.674053 5704 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 06:41:09.674112 5704 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 06:41:09.674087 5704 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.812549 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.825310 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.840548 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.853331 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.863793 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.866935 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.866974 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.866988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.867011 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.867027 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.874501 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.886965 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.899890 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.912573 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:09Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.969258 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.969290 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.969301 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.969318 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:09 crc kubenswrapper[4492]: I0220 06:41:09.969331 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:09Z","lastTransitionTime":"2026-02-20T06:41:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.071906 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.071954 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.071962 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.071982 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.071994 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.173964 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.174004 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.174018 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.174037 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.174047 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.275760 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.275797 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.275809 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.275824 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.275836 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.378043 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.378094 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.378114 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.378130 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.378140 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.481252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.481288 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.481297 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.481314 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.481324 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.527900 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 13:18:09.085998775 +0000 UTC Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.555948 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.555969 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.556013 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:10 crc kubenswrapper[4492]: E0220 06:41:10.556065 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:10 crc kubenswrapper[4492]: E0220 06:41:10.556118 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:10 crc kubenswrapper[4492]: E0220 06:41:10.556172 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.583303 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.583335 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.583348 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.583361 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.583371 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.685623 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.685657 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.685667 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.685684 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.685698 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.741550 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/1.log" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.742068 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/0.log" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.744648 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34" exitCode=1 Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.744684 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.744748 4492 scope.go:117] "RemoveContainer" containerID="8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.745172 4492 scope.go:117] "RemoveContainer" containerID="52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34" Feb 20 06:41:10 crc kubenswrapper[4492]: E0220 06:41:10.745300 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.757572 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.765066 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.773692 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.780275 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.787604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.787627 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.787637 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.787650 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.787659 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.794589 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.803412 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.810428 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.818580 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.826767 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.834196 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.845990 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a46fc58112e85ba687d736bee91126892b48d543816223c6e0f001afcd1a4aa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:09Z\\\",\\\"message\\\":\\\"220 06:41:09.672966 5704 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:09.673130 5704 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:09.673395 5704 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0220 06:41:09.673660 5704 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:09.673837 5704 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0220 06:41:09.673945 5704 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0220 06:41:09.673997 5704 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0220 06:41:09.674007 5704 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0220 06:41:09.674042 5704 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0220 06:41:09.674054 5704 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0220 06:41:09.674053 5704 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 06:41:09.674112 5704 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0220 06:41:09.674087 5704 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.854002 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.861973 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.871844 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.889216 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.889260 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.889270 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.889284 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.889294 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.991858 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.991895 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.991904 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.991917 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:10 crc kubenswrapper[4492]: I0220 06:41:10.991926 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:10Z","lastTransitionTime":"2026-02-20T06:41:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.094015 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.094049 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.094057 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.094069 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.094077 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.196655 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.196692 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.196701 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.196716 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.196727 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.298386 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.298421 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.298431 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.298444 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.298454 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.400913 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.400939 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.400947 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.400959 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.400967 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.502957 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.503154 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.503163 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.503176 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.503185 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.528251 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 00:46:38.07829246 +0000 UTC Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.604788 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.604817 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.604824 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.604836 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.604844 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.706410 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.706436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.706444 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.706454 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.706463 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.748182 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/1.log" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.750734 4492 scope.go:117] "RemoveContainer" containerID="52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34" Feb 20 06:41:11 crc kubenswrapper[4492]: E0220 06:41:11.750859 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.758900 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.765538 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.773884 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.781817 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.792809 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.801603 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.808605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.808654 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.808664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.808676 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.808684 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.809499 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.818315 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.827778 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.836551 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.849759 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.859601 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.868308 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.878391 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:11Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.910895 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.910927 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.910939 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.910953 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:11 crc kubenswrapper[4492]: I0220 06:41:11.910964 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:11Z","lastTransitionTime":"2026-02-20T06:41:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.013639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.013678 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.013688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.013703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.013714 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.115694 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.115720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.115729 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.115740 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.115747 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.217190 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.217211 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.217220 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.217230 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.217238 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.319228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.319349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.319430 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.319523 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.319607 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.421998 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.422033 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.422045 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.422060 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.422070 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.524243 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.524298 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.524310 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.524325 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.524333 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.528517 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 02:47:39.835170745 +0000 UTC Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.556923 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.557014 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.557019 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:12 crc kubenswrapper[4492]: E0220 06:41:12.557110 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:12 crc kubenswrapper[4492]: E0220 06:41:12.557161 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:12 crc kubenswrapper[4492]: E0220 06:41:12.557230 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.626772 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.626808 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.626817 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.626829 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.626842 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.728588 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.728643 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.728655 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.728670 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.728700 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.830362 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.830387 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.830401 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.830432 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.830442 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.932223 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.932249 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.932260 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.932269 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:12 crc kubenswrapper[4492]: I0220 06:41:12.932280 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:12Z","lastTransitionTime":"2026-02-20T06:41:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.034257 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.034295 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.034309 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.034326 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.034337 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.135837 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.135871 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.135880 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.135895 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.135904 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.237895 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.237948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.237961 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.237988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.238002 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.339642 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.339670 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.339684 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.339697 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.339708 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.441985 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.442018 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.442050 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.442063 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.442071 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.529100 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 01:56:34.565001323 +0000 UTC Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.543817 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.543848 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.543857 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.543870 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.543881 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.569552 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.579113 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.586919 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.594562 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.603353 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.615336 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.629553 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.637015 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.645578 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.645615 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.645627 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.645650 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.645663 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.645919 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.655136 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.664114 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.673180 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.681491 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.688932 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.747762 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.747887 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.747908 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.747923 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.747934 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.850404 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.850428 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.850437 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.850448 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.850456 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.952455 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.952535 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.952549 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.952565 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.952576 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.979407 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.979443 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.979453 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.979494 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.979507 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:13 crc kubenswrapper[4492]: E0220 06:41:13.987697 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.991461 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.991513 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.991528 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.991547 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:13 crc kubenswrapper[4492]: I0220 06:41:13.991560 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:13Z","lastTransitionTime":"2026-02-20T06:41:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.001744 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.004369 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.004395 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.004426 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.004438 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.004448 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.014208 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.016892 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.016928 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.016938 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.016954 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.016968 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.026426 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.029493 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.029612 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.029684 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.029751 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.029813 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.039944 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.040051 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.054256 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.054358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.054418 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.054501 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.054560 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.154149 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl"] Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.154742 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.156832 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.160719 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.161182 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.161233 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.161246 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.161259 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.161275 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.168227 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.178111 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.188865 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.196803 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.208373 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.216535 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.221419 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/921784a1-2034-4ca6-b954-af969bfefd1a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.221567 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/921784a1-2034-4ca6-b954-af969bfefd1a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.221742 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q64zj\" (UniqueName: \"kubernetes.io/projected/921784a1-2034-4ca6-b954-af969bfefd1a-kube-api-access-q64zj\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.221807 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/921784a1-2034-4ca6-b954-af969bfefd1a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.226411 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.234964 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.244685 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.254654 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.263327 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.263360 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.263373 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.263391 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.263407 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.265622 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.274202 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.281274 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.289641 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.296510 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.322923 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q64zj\" (UniqueName: \"kubernetes.io/projected/921784a1-2034-4ca6-b954-af969bfefd1a-kube-api-access-q64zj\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.322992 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/921784a1-2034-4ca6-b954-af969bfefd1a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.323032 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/921784a1-2034-4ca6-b954-af969bfefd1a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.323071 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/921784a1-2034-4ca6-b954-af969bfefd1a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.323678 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/921784a1-2034-4ca6-b954-af969bfefd1a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.323894 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/921784a1-2034-4ca6-b954-af969bfefd1a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.333445 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/921784a1-2034-4ca6-b954-af969bfefd1a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.343804 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q64zj\" (UniqueName: \"kubernetes.io/projected/921784a1-2034-4ca6-b954-af969bfefd1a-kube-api-access-q64zj\") pod \"ovnkube-control-plane-749d76644c-c6ffl\" (UID: \"921784a1-2034-4ca6-b954-af969bfefd1a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.366664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.366693 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.366704 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.366720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.366732 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.469058 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.469096 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.469107 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.469122 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.469136 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.470239 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" Feb 20 06:41:14 crc kubenswrapper[4492]: W0220 06:41:14.482972 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod921784a1_2034_4ca6_b954_af969bfefd1a.slice/crio-3f313c30bc3f17e5c8bf96fb78a5dab3638b986932a9c95cbc6a0e1af9a61c00 WatchSource:0}: Error finding container 3f313c30bc3f17e5c8bf96fb78a5dab3638b986932a9c95cbc6a0e1af9a61c00: Status 404 returned error can't find the container with id 3f313c30bc3f17e5c8bf96fb78a5dab3638b986932a9c95cbc6a0e1af9a61c00 Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.530569 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 18:54:06.048814512 +0000 UTC Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.556844 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.556976 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.557129 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.557288 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.557299 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.557524 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.571660 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.571706 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.571721 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.571746 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.571762 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.673465 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.673658 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.673719 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.673833 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.673891 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.761851 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" event={"ID":"921784a1-2034-4ca6-b954-af969bfefd1a","Type":"ContainerStarted","Data":"58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.762198 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" event={"ID":"921784a1-2034-4ca6-b954-af969bfefd1a","Type":"ContainerStarted","Data":"dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.762282 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" event={"ID":"921784a1-2034-4ca6-b954-af969bfefd1a","Type":"ContainerStarted","Data":"3f313c30bc3f17e5c8bf96fb78a5dab3638b986932a9c95cbc6a0e1af9a61c00"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.775559 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.775586 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.775595 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.775615 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.775625 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.779865 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.794341 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.807173 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.816361 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.825185 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.834393 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.844068 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.851931 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.856824 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-p9vbm"] Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.857309 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:14 crc kubenswrapper[4492]: E0220 06:41:14.857429 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.863079 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.870970 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.877822 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.877864 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.877879 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.877899 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.877943 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.880852 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.890130 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.898902 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.906412 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.913909 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.922063 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.928888 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.928927 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dn6d\" (UniqueName: \"kubernetes.io/projected/0825f068-2c78-44a0-8f24-679e4521f069-kube-api-access-4dn6d\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.929890 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.940523 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.947675 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.956965 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.965738 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.974261 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.980537 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.980570 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.980583 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.980611 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.980625 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:14Z","lastTransitionTime":"2026-02-20T06:41:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.981861 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:14 crc kubenswrapper[4492]: I0220 06:41:14.990255 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.000713 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:14Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.009056 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.023044 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.030038 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.030081 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dn6d\" (UniqueName: \"kubernetes.io/projected/0825f068-2c78-44a0-8f24-679e4521f069-kube-api-access-4dn6d\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:15 crc kubenswrapper[4492]: E0220 06:41:15.030286 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:15 crc kubenswrapper[4492]: E0220 06:41:15.030422 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:15.530401452 +0000 UTC m=+32.301690430 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.032142 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.043575 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dn6d\" (UniqueName: \"kubernetes.io/projected/0825f068-2c78-44a0-8f24-679e4521f069-kube-api-access-4dn6d\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.049416 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.082252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.082292 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.082302 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.082320 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.082330 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.100327 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.108132 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.184322 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.184369 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.184380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.184401 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.184413 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.286437 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.286491 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.286502 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.286519 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.286532 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.389241 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.389436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.389446 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.389465 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.389498 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.492369 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.492417 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.492433 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.492463 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.492494 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.531028 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 23:09:15.71703662 +0000 UTC Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.535556 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:15 crc kubenswrapper[4492]: E0220 06:41:15.535681 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:15 crc kubenswrapper[4492]: E0220 06:41:15.535739 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:16.535724624 +0000 UTC m=+33.307013601 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.595744 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.595778 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.595789 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.595803 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.595817 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.698705 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.698756 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.698767 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.698790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.698802 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.800382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.800411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.800423 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.800438 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.800451 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.901931 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.901958 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.901966 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.901978 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:15 crc kubenswrapper[4492]: I0220 06:41:15.901989 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:15Z","lastTransitionTime":"2026-02-20T06:41:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.004268 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.004297 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.004307 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.004322 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.004333 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.106501 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.106536 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.106554 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.106569 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.106579 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.208761 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.208800 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.208811 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.208841 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.208853 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.310541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.310575 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.310585 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.310601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.310613 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.343143 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.343252 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:41:32.343234389 +0000 UTC m=+49.114523367 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.412665 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.412688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.412697 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.412708 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.412717 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.443532 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.443580 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.443603 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.443620 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443692 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443711 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443741 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:32.443730225 +0000 UTC m=+49.215019203 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443767 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:32.443753889 +0000 UTC m=+49.215042867 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443775 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443795 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443810 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443853 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:32.44384373 +0000 UTC m=+49.215132709 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443879 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443927 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.443945 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.444050 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:32.444021718 +0000 UTC m=+49.215310706 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.468129 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.480974 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.490744 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.498233 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.506332 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.513954 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.513984 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.513993 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.514004 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.514012 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.515005 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.526175 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.531456 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 19:19:21.751208287 +0000 UTC Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.536174 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.544777 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.544887 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.544942 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:18.544929997 +0000 UTC m=+35.316218985 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.550389 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.556333 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.556360 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.556390 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.556417 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.556432 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.556570 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.556620 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:16 crc kubenswrapper[4492]: E0220 06:41:16.556701 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.559393 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.575834 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.586398 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.593887 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.601886 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.608832 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.616801 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.616903 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.616980 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.617041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.617104 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.618720 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.626402 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:16Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.720435 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.720503 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.720524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.720548 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.720562 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.822366 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.822405 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.822416 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.822430 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.822442 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.925033 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.925073 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.925082 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.925104 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:16 crc kubenswrapper[4492]: I0220 06:41:16.925118 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:16Z","lastTransitionTime":"2026-02-20T06:41:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.027211 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.027248 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.027262 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.027282 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.027295 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.129905 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.129966 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.129976 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.129992 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.130003 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.231769 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.231797 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.231815 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.231826 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.231835 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.333929 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.333949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.333957 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.333966 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.333973 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.435985 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.436015 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.436026 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.436037 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.436046 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.532037 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 02:16:17.739014202 +0000 UTC Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.537603 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.537629 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.537652 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.537665 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.537673 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.640331 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.640380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.640391 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.640411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.640430 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.742374 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.742399 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.742408 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.742420 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.742429 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.844805 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.844922 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.844977 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.845035 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.845090 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.947262 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.947295 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.947304 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.947317 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:17 crc kubenswrapper[4492]: I0220 06:41:17.947327 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:17Z","lastTransitionTime":"2026-02-20T06:41:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.048981 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.049013 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.049043 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.049057 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.049066 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.150826 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.150875 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.150885 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.150894 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.150901 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.252588 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.252625 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.252634 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.252649 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.252677 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.354867 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.354906 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.354918 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.354935 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.354947 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.456761 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.456790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.456800 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.456810 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.456820 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.532333 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 19:54:51.837778841 +0000 UTC Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.556668 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.556742 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.556716 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.556717 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:18 crc kubenswrapper[4492]: E0220 06:41:18.556897 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:18 crc kubenswrapper[4492]: E0220 06:41:18.557005 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:18 crc kubenswrapper[4492]: E0220 06:41:18.557057 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:18 crc kubenswrapper[4492]: E0220 06:41:18.557118 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.558239 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.558262 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.558270 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.558280 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.558288 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.563044 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:18 crc kubenswrapper[4492]: E0220 06:41:18.563152 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:18 crc kubenswrapper[4492]: E0220 06:41:18.563202 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:22.563189579 +0000 UTC m=+39.334478557 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.660255 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.660394 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.660411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.660424 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.660433 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.762463 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.762529 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.762540 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.762553 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.762561 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.864405 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.864434 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.864442 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.864454 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.864462 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.966755 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.966789 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.966799 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.966810 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:18 crc kubenswrapper[4492]: I0220 06:41:18.966818 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:18Z","lastTransitionTime":"2026-02-20T06:41:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.068856 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.068896 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.068905 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.068913 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.068920 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.170943 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.170974 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.170987 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.171021 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.171030 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.273082 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.273117 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.273129 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.273140 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.273151 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.375265 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.375330 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.375341 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.375369 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.375382 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.477499 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.477532 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.477544 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.477574 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.477583 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.533015 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 19:45:14.061041021 +0000 UTC Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.579664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.579697 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.579705 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.579734 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.579742 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.681957 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.681988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.681997 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.682013 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.682024 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.784442 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.784492 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.784502 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.784514 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.784522 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.886617 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.886645 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.886674 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.886696 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.886709 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.988435 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.988467 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.988493 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.988505 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:19 crc kubenswrapper[4492]: I0220 06:41:19.988514 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:19Z","lastTransitionTime":"2026-02-20T06:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.090953 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.090985 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.090994 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.091023 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.091033 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.193191 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.193227 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.193237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.193250 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.193262 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.295142 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.295167 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.295177 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.295188 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.295196 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.397509 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.397571 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.397620 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.397669 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.397703 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.499828 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.499866 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.499878 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.499893 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.499904 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.533990 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 18:18:27.328238103 +0000 UTC Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.556323 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.556344 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.556350 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.556330 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:20 crc kubenswrapper[4492]: E0220 06:41:20.556451 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:20 crc kubenswrapper[4492]: E0220 06:41:20.556544 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:20 crc kubenswrapper[4492]: E0220 06:41:20.556673 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:20 crc kubenswrapper[4492]: E0220 06:41:20.556731 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.601258 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.601292 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.601304 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.601318 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.601330 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.702663 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.702721 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.702735 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.702750 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.702760 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.804842 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.804870 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.804897 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.804909 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.804919 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.906614 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.906643 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.906661 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.906672 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:20 crc kubenswrapper[4492]: I0220 06:41:20.906684 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:20Z","lastTransitionTime":"2026-02-20T06:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.008358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.008381 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.008389 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.008401 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.008430 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.110561 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.110589 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.110597 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.110609 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.110618 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.212505 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.212539 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.212550 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.212563 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.212572 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.314866 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.314928 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.314940 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.314965 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.314978 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.417395 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.417446 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.417456 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.417469 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.417496 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.519020 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.519070 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.519080 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.519101 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.519114 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.534352 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 21:54:12.299543502 +0000 UTC Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.621188 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.621221 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.621232 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.621242 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.621252 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.723749 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.723784 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.723795 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.723811 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.723822 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.825983 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.826029 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.826041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.826061 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.826075 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.927920 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.927968 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.927980 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.927997 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:21 crc kubenswrapper[4492]: I0220 06:41:21.928010 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:21Z","lastTransitionTime":"2026-02-20T06:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.029663 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.029702 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.029721 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.029736 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.029750 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.120818 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.121454 4492 scope.go:117] "RemoveContainer" containerID="52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.132140 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.132184 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.132194 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.132209 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.132220 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.241097 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.241402 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.241413 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.241429 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.241451 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.344781 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.344835 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.344845 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.344868 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.344881 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.446994 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.447043 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.447054 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.447071 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.447081 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.534653 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 01:19:58.587087225 +0000 UTC Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.549868 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.549903 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.549911 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.549926 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.549936 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.556253 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.556324 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.556253 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:22 crc kubenswrapper[4492]: E0220 06:41:22.556357 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.556324 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:22 crc kubenswrapper[4492]: E0220 06:41:22.556443 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:22 crc kubenswrapper[4492]: E0220 06:41:22.556534 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:22 crc kubenswrapper[4492]: E0220 06:41:22.556590 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.601758 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:22 crc kubenswrapper[4492]: E0220 06:41:22.601924 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:22 crc kubenswrapper[4492]: E0220 06:41:22.601978 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:30.601961841 +0000 UTC m=+47.373250819 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.652441 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.652489 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.652500 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.652516 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.652525 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.754553 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.754594 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.754605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.754624 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.754635 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.786048 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/1.log" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.789054 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.789434 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.801761 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.818918 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.828395 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.840123 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.852174 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.856707 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.856742 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.856753 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.856770 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.856780 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.860956 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.869759 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.879598 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.887947 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.897145 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.905969 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.915206 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.923136 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.930996 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.941839 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.952522 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:22Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.959311 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.959342 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.959352 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.959367 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:22 crc kubenswrapper[4492]: I0220 06:41:22.959377 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:22Z","lastTransitionTime":"2026-02-20T06:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.061783 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.061836 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.061847 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.061862 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.061872 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.164221 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.164267 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.164277 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.164300 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.164312 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.266107 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.266128 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.266137 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.266147 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.266157 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.368499 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.368525 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.368536 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.368546 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.368556 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.471074 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.471101 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.471110 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.471121 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.471128 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.535372 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 15:27:18.183300063 +0000 UTC Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.568804 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.572576 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.572604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.572614 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.572627 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.572637 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.578408 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.588035 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.596700 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.605578 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.616041 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.629947 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.637829 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.646796 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.656408 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.664156 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.672722 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.674005 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.674042 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.674052 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.674064 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.674073 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.683155 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.695372 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.707403 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.726509 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.776179 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.776217 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.776226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.776243 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.776254 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.793104 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/2.log" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.793685 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/1.log" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.796420 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114" exitCode=1 Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.796460 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.796516 4492 scope.go:117] "RemoveContainer" containerID="52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.797678 4492 scope.go:117] "RemoveContainer" containerID="f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114" Feb 20 06:41:23 crc kubenswrapper[4492]: E0220 06:41:23.797972 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.809223 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.819793 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.828591 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.837760 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.848648 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.858576 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.868021 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.877106 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.878676 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.878714 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.878751 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.878774 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.878789 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.888751 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.898618 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.913370 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52f9c1d185cef26982fa99dee9ef988b4de44b91a8252a45770bd8ce00396c34\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:10Z\\\",\\\"message\\\":\\\"pired or is not yet valid: current time 2026-02-20T06:41:10Z is after 2025-08-24T17:21:41Z]\\\\nI0220 06:41:10.419642 5826 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI0220 06:41:10.419640 5826 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-console/networking-console-plugin\\\\\\\"}\\\\nI0220 06:41:10.419623 5826 services_controller.go:434] Service openshift-ingress-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-ingress-operator 06971ebf-e383-4ded-a48b-3b2e61be789e 4180 0 2025-02-23 05:12:34 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:ingress-operator] map[capability.openshift.io/name:Ingress include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc0006afa57 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServiceP\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:09Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.922808 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.936891 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.949060 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.960657 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.968575 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:23Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.981169 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.981206 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.981219 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.981238 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:23 crc kubenswrapper[4492]: I0220 06:41:23.981250 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:23Z","lastTransitionTime":"2026-02-20T06:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.083882 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.083923 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.083933 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.083949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.083960 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.186047 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.186187 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.186249 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.186338 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.186401 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.288881 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.288933 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.288947 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.288972 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.288990 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.367189 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.367237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.367251 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.367269 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.367283 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.377499 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.380902 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.380952 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.380963 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.380985 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.381003 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.389683 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.392575 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.392609 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.392641 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.392668 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.392680 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.401114 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.403618 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.403682 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.403699 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.403714 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.403723 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.413979 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.416706 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.416765 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.416778 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.416791 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.416801 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.425133 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.425238 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.426440 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.426488 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.426500 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.426515 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.426523 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.528930 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.529006 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.529023 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.529047 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.529064 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.536455 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 15:08:24.041706234 +0000 UTC Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.556822 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.556910 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.556916 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.556825 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.556988 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.557133 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.557204 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.557289 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.631374 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.631406 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.631420 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.631436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.631450 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.733377 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.733411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.733427 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.733440 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.733452 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.801744 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/2.log" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.805586 4492 scope.go:117] "RemoveContainer" containerID="f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114" Feb 20 06:41:24 crc kubenswrapper[4492]: E0220 06:41:24.805787 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.816661 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.824942 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.834347 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.835279 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.835321 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.835337 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.835367 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.835384 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.842598 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.852000 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.861019 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.870128 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.880969 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.891322 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.899772 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.913165 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.923237 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.936261 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.937344 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.937377 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.937387 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.937404 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.937415 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:24Z","lastTransitionTime":"2026-02-20T06:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.949306 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.957457 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:24 crc kubenswrapper[4492]: I0220 06:41:24.967749 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:24Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.039931 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.039965 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.039975 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.039991 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.040002 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.142115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.142151 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.142162 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.142177 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.142200 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.244383 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.244409 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.244438 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.244451 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.244458 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.345999 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.346028 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.346038 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.346052 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.346060 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.447908 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.447961 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.447973 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.447989 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.447997 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.537212 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 03:43:42.946959579 +0000 UTC Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.549927 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.550083 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.550191 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.550278 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.550347 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.652324 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.653207 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.653246 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.653282 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.653303 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.755894 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.755934 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.755943 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.755961 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.755973 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.858960 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.859009 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.859022 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.859041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.859055 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.961295 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.961330 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.961342 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.961356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:25 crc kubenswrapper[4492]: I0220 06:41:25.961366 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:25Z","lastTransitionTime":"2026-02-20T06:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.063577 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.063611 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.063622 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.063636 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.063646 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.165645 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.165714 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.165728 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.165746 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.165784 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.267608 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.267655 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.267668 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.267684 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.267696 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.369980 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.370013 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.370023 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.370058 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.370069 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.471771 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.471906 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.471979 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.472045 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.472110 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.538663 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 03:30:12.502925843 +0000 UTC Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.555913 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.556017 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.555948 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.556142 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:26 crc kubenswrapper[4492]: E0220 06:41:26.556336 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:26 crc kubenswrapper[4492]: E0220 06:41:26.556707 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:26 crc kubenswrapper[4492]: E0220 06:41:26.556587 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:26 crc kubenswrapper[4492]: E0220 06:41:26.556832 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.574340 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.574370 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.574380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.574399 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.574414 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.676867 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.676917 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.676929 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.676939 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.676948 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.778568 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.778600 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.778610 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.778624 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.778637 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.880237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.880285 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.880297 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.880308 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.880316 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.981948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.981976 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.981987 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.982001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:26 crc kubenswrapper[4492]: I0220 06:41:26.982013 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:26Z","lastTransitionTime":"2026-02-20T06:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.083664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.083698 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.083709 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.083724 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.083734 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.185450 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.185497 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.185513 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.185528 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.185539 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.287436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.287465 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.287501 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.287519 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.287529 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.393169 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.393207 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.393219 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.393234 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.393251 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.495394 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.495430 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.495438 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.495452 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.495463 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.538768 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 23:31:06.790042924 +0000 UTC Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.597715 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.597755 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.597766 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.597780 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.597802 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.699465 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.699604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.699685 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.699753 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.699833 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.801853 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.801985 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.802051 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.802118 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.802177 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.903885 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.903919 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.903928 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.903941 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:27 crc kubenswrapper[4492]: I0220 06:41:27.903952 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:27Z","lastTransitionTime":"2026-02-20T06:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.006080 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.006115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.006124 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.006136 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.006144 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.108086 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.108117 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.108130 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.108143 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.108152 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.210048 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.210088 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.210099 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.210113 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.210122 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.312066 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.312194 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.312265 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.312353 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.312433 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.414025 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.414053 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.414063 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.414073 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.414082 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.515781 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.515829 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.515840 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.515857 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.515868 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.539544 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 07:18:15.780733512 +0000 UTC Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.555907 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.555958 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:28 crc kubenswrapper[4492]: E0220 06:41:28.556017 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.556031 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.556072 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:28 crc kubenswrapper[4492]: E0220 06:41:28.556104 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:28 crc kubenswrapper[4492]: E0220 06:41:28.556165 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:28 crc kubenswrapper[4492]: E0220 06:41:28.556224 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.617973 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.618003 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.618013 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.618026 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.618037 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.720189 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.720229 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.720239 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.720259 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.720271 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.822001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.822039 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.822053 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.822066 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.822078 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.923988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.924017 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.924025 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.924042 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:28 crc kubenswrapper[4492]: I0220 06:41:28.924054 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:28Z","lastTransitionTime":"2026-02-20T06:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.025621 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.025661 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.025672 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.025686 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.025698 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.127171 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.127199 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.127213 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.127227 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.127236 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.229448 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.229528 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.229551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.229568 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.229580 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.331430 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.331459 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.331494 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.331507 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.331516 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.433616 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.433647 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.433659 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.433670 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.433682 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.535234 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.535273 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.535282 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.535298 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.535310 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.540500 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 18:48:43.162386194 +0000 UTC Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.636814 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.637094 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.637168 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.637238 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.637299 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.739147 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.739175 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.739184 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.739194 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.739201 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.841231 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.841265 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.841276 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.841291 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.841301 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.942794 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.942816 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.942824 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.942841 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:29 crc kubenswrapper[4492]: I0220 06:41:29.942849 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:29Z","lastTransitionTime":"2026-02-20T06:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.044992 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.045039 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.045052 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.045064 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.045073 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.147435 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.147507 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.147517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.147533 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.147544 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.249237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.249262 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.249271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.249283 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.249291 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.351116 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.351143 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.351152 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.351163 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.351173 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.452339 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.452530 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.452605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.452670 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.452728 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.541432 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 13:22:40.205708598 +0000 UTC Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.554994 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.555021 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.555050 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.555064 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.555073 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.556183 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:30 crc kubenswrapper[4492]: E0220 06:41:30.556269 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.556371 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.556377 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.556514 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:30 crc kubenswrapper[4492]: E0220 06:41:30.556606 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:30 crc kubenswrapper[4492]: E0220 06:41:30.556742 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:30 crc kubenswrapper[4492]: E0220 06:41:30.556860 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.657544 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.657581 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.657590 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.657603 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.657613 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.672401 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:30 crc kubenswrapper[4492]: E0220 06:41:30.672577 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:30 crc kubenswrapper[4492]: E0220 06:41:30.672651 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:41:46.672633113 +0000 UTC m=+63.443922092 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.759511 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.759581 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.759591 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.759607 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.759618 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.861504 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.861531 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.861556 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.861566 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.861573 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.963313 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.963346 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.963354 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.963366 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:30 crc kubenswrapper[4492]: I0220 06:41:30.963377 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:30Z","lastTransitionTime":"2026-02-20T06:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.065565 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.065597 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.065605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.065619 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.065628 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.167553 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.167579 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.167588 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.167600 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.167610 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.269654 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.269685 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.269694 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.269705 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.269714 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.371014 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.371043 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.371051 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.371060 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.371068 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.472751 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.472780 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.472788 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.472797 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.472805 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.542457 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 21:17:03.314658401 +0000 UTC Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.556938 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:31 crc kubenswrapper[4492]: E0220 06:41:31.557043 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.574305 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.574360 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.574371 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.574382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.574390 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.676317 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.676351 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.676359 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.676373 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.676381 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.778427 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.778594 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.778666 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.778728 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.778788 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.880307 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.880334 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.880342 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.880354 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.880364 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.981984 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.982133 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.982214 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.982274 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:31 crc kubenswrapper[4492]: I0220 06:41:31.982326 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:31Z","lastTransitionTime":"2026-02-20T06:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.084298 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.084332 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.084342 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.084352 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.084364 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.185965 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.186001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.186009 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.186022 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.186030 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.288216 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.288254 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.288265 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.288278 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.288287 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.382845 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.383005 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:42:04.382984103 +0000 UTC m=+81.154273091 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.389950 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.389979 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.389987 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.389998 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.390006 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.483910 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.483946 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.483966 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.483985 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484065 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484064 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484102 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:42:04.484090008 +0000 UTC m=+81.255378986 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484109 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484128 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484150 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484154 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484170 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484180 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484131 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:42:04.484114324 +0000 UTC m=+81.255403303 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484204 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:42:04.484192654 +0000 UTC m=+81.255481642 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.484233 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:42:04.484224875 +0000 UTC m=+81.255513843 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.491631 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.491658 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.491667 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.491682 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.491689 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.543176 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 22:03:27.967778369 +0000 UTC Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.556471 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.556506 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.556575 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.556492 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.556648 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:32 crc kubenswrapper[4492]: E0220 06:41:32.556697 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.593382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.593417 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.593425 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.593434 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.593442 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.695061 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.695086 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.695096 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.695107 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.695118 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.796215 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.796242 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.796252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.796262 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.796271 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.897697 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.897752 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.897763 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.897772 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.897779 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.999140 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.999158 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.999165 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.999175 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:32 crc kubenswrapper[4492]: I0220 06:41:32.999182 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:32Z","lastTransitionTime":"2026-02-20T06:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.100688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.100720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.100729 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.100740 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.100749 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.202293 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.202327 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.202338 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.202351 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.202360 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.303819 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.303839 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.303848 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.303857 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.303866 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.405463 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.405505 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.405515 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.405524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.405531 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.507455 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.507495 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.507504 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.507514 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.507521 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.544170 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 22:34:43.941255913 +0000 UTC Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.556566 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:33 crc kubenswrapper[4492]: E0220 06:41:33.556666 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.567194 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.575567 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.582468 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.589744 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.598703 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.606978 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.609054 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.609076 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.609085 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.609094 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.609101 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.624253 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.632397 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.640783 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.651257 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.661303 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.668623 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.675580 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.684213 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.690801 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.698763 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:33Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.710622 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.710728 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.710795 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.710872 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.710940 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.813074 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.813101 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.813111 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.813122 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.813130 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.914551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.914598 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.914612 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.914630 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:33 crc kubenswrapper[4492]: I0220 06:41:33.914643 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:33Z","lastTransitionTime":"2026-02-20T06:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.016660 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.016690 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.016698 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.016709 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.016717 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.118902 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.118931 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.118940 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.118952 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.118961 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.220668 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.220688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.220696 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.220704 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.220710 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.322084 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.322113 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.322120 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.322129 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.322136 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.423663 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.423754 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.423769 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.423786 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.423798 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.525315 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.525341 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.525349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.525363 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.525371 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.545089 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 10:13:23.607383597 +0000 UTC Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.556352 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.556409 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.556442 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.556452 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.556523 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.556591 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.627228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.627252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.627261 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.627271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.627279 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.728969 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.728999 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729008 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729018 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729026 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729704 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729731 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729740 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729748 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.729754 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.738591 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:34Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.740697 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.740720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.740728 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.740738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.740745 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.749051 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:34Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.751513 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.751549 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.751560 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.751580 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.751587 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.761662 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:34Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.763813 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.763842 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.763852 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.763862 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.763869 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.771994 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:34Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.774062 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.774089 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.774097 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.774109 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.774118 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.781617 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:34Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:34 crc kubenswrapper[4492]: E0220 06:41:34.781756 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.830504 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.830530 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.830538 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.830549 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.830556 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.932310 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.932422 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.932504 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.932564 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:34 crc kubenswrapper[4492]: I0220 06:41:34.932634 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:34Z","lastTransitionTime":"2026-02-20T06:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.034541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.034563 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.034572 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.034582 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.034590 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.136029 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.136060 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.136070 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.136081 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.136090 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.241059 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.241085 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.241102 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.241112 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.241121 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.342783 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.342813 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.342821 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.342834 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.342842 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.444567 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.444593 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.444602 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.444611 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.444619 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.545235 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 22:55:27.900165106 +0000 UTC Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.546096 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.546117 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.546126 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.546137 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.546145 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.556591 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:35 crc kubenswrapper[4492]: E0220 06:41:35.556672 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.648425 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.648452 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.648459 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.648469 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.648493 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.749803 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.749825 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.749833 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.749845 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.749852 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.851671 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.851700 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.851709 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.851721 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.851729 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.953739 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.953760 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.953767 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.953777 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:35 crc kubenswrapper[4492]: I0220 06:41:35.953784 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:35Z","lastTransitionTime":"2026-02-20T06:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.055664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.055685 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.055692 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.055701 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.055709 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.160226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.160259 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.160267 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.160278 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.160285 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.261636 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.261657 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.261665 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.261674 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.261682 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.362813 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.362850 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.362860 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.362876 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.362885 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.464949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.464975 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.464983 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.464992 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.465000 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.545418 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 23:17:41.826630301 +0000 UTC Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.556803 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.556854 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.556819 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:36 crc kubenswrapper[4492]: E0220 06:41:36.556947 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:36 crc kubenswrapper[4492]: E0220 06:41:36.557049 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:36 crc kubenswrapper[4492]: E0220 06:41:36.557146 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.566213 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.566318 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.566389 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.566448 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.566532 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.668275 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.668301 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.668318 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.668331 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.668340 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.770108 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.770135 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.770143 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.770153 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.770161 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.871469 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.871521 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.871530 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.871540 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.871547 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.973813 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.973836 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.973845 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.973854 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:36 crc kubenswrapper[4492]: I0220 06:41:36.973862 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:36Z","lastTransitionTime":"2026-02-20T06:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.075632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.075774 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.075849 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.075926 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.075999 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.177395 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.177423 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.177446 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.177460 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.177469 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.279630 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.279656 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.279664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.279672 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.279679 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.381271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.381299 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.381309 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.381321 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.381329 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.483159 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.483186 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.483194 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.483203 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.483209 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.546259 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 01:05:33.438607688 +0000 UTC Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.556569 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:37 crc kubenswrapper[4492]: E0220 06:41:37.556655 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.567975 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.574888 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.578270 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.585237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.585260 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.585268 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.585278 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.585286 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.588411 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.596731 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.610003 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.618494 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.625735 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.634517 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.640497 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.647439 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.654007 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.661764 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.667958 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.676722 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.684698 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.686855 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.686878 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.686887 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.686920 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.686929 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.692003 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.699511 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:37Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.788364 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.788390 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.788399 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.788410 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.788418 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.890694 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.890730 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.890737 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.890753 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.890762 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.992530 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.992557 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.992565 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.992578 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:37 crc kubenswrapper[4492]: I0220 06:41:37.992603 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:37Z","lastTransitionTime":"2026-02-20T06:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.094271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.094320 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.094334 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.094354 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.094367 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.196358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.196400 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.196410 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.196433 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.196445 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.297983 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.298016 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.298024 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.298033 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.298039 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.399445 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.399522 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.399531 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.399545 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.399554 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.501923 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.501951 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.501966 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.501975 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.501982 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.546708 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 19:49:37.701218217 +0000 UTC Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.555999 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.556045 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.556311 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:38 crc kubenswrapper[4492]: E0220 06:41:38.556371 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:38 crc kubenswrapper[4492]: E0220 06:41:38.556416 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:38 crc kubenswrapper[4492]: E0220 06:41:38.556496 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.556628 4492 scope.go:117] "RemoveContainer" containerID="f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114" Feb 20 06:41:38 crc kubenswrapper[4492]: E0220 06:41:38.556772 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.603231 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.603264 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.603272 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.603283 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.603292 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.704940 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.705068 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.705145 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.705218 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.705272 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.806658 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.806691 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.806700 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.806714 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.806724 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.908058 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.908193 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.908251 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.908311 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:38 crc kubenswrapper[4492]: I0220 06:41:38.908376 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:38Z","lastTransitionTime":"2026-02-20T06:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.010158 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.010175 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.010184 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.010196 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.010204 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.111914 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.111946 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.111956 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.111986 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.112002 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.213600 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.213629 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.213638 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.213649 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.213658 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.315300 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.315321 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.315329 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.315338 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.315346 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.417136 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.417160 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.417170 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.417179 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.417187 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.518205 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.518238 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.518249 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.518262 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.518271 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.547093 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 18:53:28.818998704 +0000 UTC Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.556428 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:39 crc kubenswrapper[4492]: E0220 06:41:39.556542 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.619341 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.619397 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.619408 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.619418 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.619426 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.721502 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.721656 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.721716 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.721774 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.721831 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.823329 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.823359 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.823368 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.823380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.823389 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.925241 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.925385 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.925459 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.925545 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:39 crc kubenswrapper[4492]: I0220 06:41:39.925606 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:39Z","lastTransitionTime":"2026-02-20T06:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.027024 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.027050 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.027058 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.027066 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.027073 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.128494 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.128523 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.128537 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.128547 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.128556 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.230587 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.230725 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.230805 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.230890 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.230964 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.332172 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.332280 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.332353 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.332416 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.332495 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.433661 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.433832 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.433901 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.433960 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.434023 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.535351 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.535377 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.535384 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.535393 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.535400 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.547875 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 04:52:53.288951593 +0000 UTC Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.556087 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.556101 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:40 crc kubenswrapper[4492]: E0220 06:41:40.556174 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.556233 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:40 crc kubenswrapper[4492]: E0220 06:41:40.556329 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:40 crc kubenswrapper[4492]: E0220 06:41:40.556268 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.636653 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.636796 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.636856 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.636906 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.636956 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.738092 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.738119 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.738128 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.738139 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.738145 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.840221 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.840376 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.840457 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.840536 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.840585 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.941712 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.941735 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.941743 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.941753 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:40 crc kubenswrapper[4492]: I0220 06:41:40.941761 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:40Z","lastTransitionTime":"2026-02-20T06:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.043568 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.043593 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.043601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.043610 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.043617 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.145658 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.145685 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.145693 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.145703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.145711 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.248372 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.248393 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.248401 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.248411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.248418 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.349962 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.350021 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.350033 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.350045 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.350054 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.451781 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.451814 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.451822 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.451832 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.451840 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.548717 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 18:49:09.804051632 +0000 UTC Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.552996 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.553039 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.553049 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.553061 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.553069 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.556326 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:41 crc kubenswrapper[4492]: E0220 06:41:41.556448 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.654555 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.654590 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.654602 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.654615 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.654623 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.756166 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.756206 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.756215 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.756228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.756237 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.858436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.858489 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.858499 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.858512 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.858520 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.960334 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.960368 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.960377 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.960390 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:41 crc kubenswrapper[4492]: I0220 06:41:41.960399 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:41Z","lastTransitionTime":"2026-02-20T06:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.061926 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.061953 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.061962 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.061975 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.061987 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.165186 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.165233 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.165243 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.165257 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.165267 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.267311 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.267339 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.267348 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.267360 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.267369 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.369331 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.369366 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.369374 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.369387 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.369395 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.471355 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.471403 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.471412 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.471424 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.471434 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.549067 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 23:50:25.382803644 +0000 UTC Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.556362 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.556408 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.556415 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:42 crc kubenswrapper[4492]: E0220 06:41:42.556494 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:42 crc kubenswrapper[4492]: E0220 06:41:42.556545 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:42 crc kubenswrapper[4492]: E0220 06:41:42.556600 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.573536 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.573564 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.573572 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.573584 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.573592 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.674906 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.674937 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.674946 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.674958 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.674968 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.776445 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.776498 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.776510 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.776520 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.776527 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.878358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.878384 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.878391 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.878401 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.878409 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.980439 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.980563 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.980630 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.980697 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:42 crc kubenswrapper[4492]: I0220 06:41:42.980760 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:42Z","lastTransitionTime":"2026-02-20T06:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.082840 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.082866 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.082917 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.082928 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.082935 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.185139 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.185175 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.185183 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.185192 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.185199 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.287143 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.287162 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.287171 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.287179 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.287188 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.389158 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.389192 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.389200 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.389209 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.389216 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.492515 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.492662 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.492762 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.492847 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.492930 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.549332 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 01:24:32.200638468 +0000 UTC Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.557377 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:43 crc kubenswrapper[4492]: E0220 06:41:43.557564 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.566585 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.575402 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.584212 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.591175 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.594349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.594377 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.594388 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.594402 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.594411 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.604093 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.612836 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.624103 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.632090 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.640553 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.647595 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.655844 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.663246 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.673360 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.680891 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.690910 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.696547 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.696581 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.696591 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.696602 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.696610 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.699945 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.706889 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:43Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.798165 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.798194 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.798202 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.798213 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.798222 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.900083 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.900228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.900317 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.900398 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:43 crc kubenswrapper[4492]: I0220 06:41:43.900488 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:43Z","lastTransitionTime":"2026-02-20T06:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.002359 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.002404 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.002413 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.002422 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.002430 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.104415 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.104454 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.104464 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.104509 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.104519 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.205826 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.205855 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.205862 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.205874 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.205881 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.308366 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.308524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.308536 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.308549 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.308557 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.410285 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.410317 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.410326 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.410340 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.410348 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.512544 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.512569 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.512577 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.512588 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.512595 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.549888 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 16:29:52.17610081 +0000 UTC Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.556131 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.556147 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:44 crc kubenswrapper[4492]: E0220 06:41:44.556215 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.556244 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:44 crc kubenswrapper[4492]: E0220 06:41:44.556283 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:44 crc kubenswrapper[4492]: E0220 06:41:44.556328 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.614628 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.614779 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.614850 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.614910 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.614970 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.716382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.716413 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.716424 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.716437 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.716447 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.817672 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.817701 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.817710 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.817723 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.817733 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.919080 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.919109 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.919118 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.919129 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:44 crc kubenswrapper[4492]: I0220 06:41:44.919138 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:44Z","lastTransitionTime":"2026-02-20T06:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.020361 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.020385 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.020394 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.020403 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.020411 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.026446 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.026482 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.026492 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.026502 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.026510 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: E0220 06:41:45.035976 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:45Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.038004 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.038025 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.038034 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.038045 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.038063 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: E0220 06:41:45.046361 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:45Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.048521 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.048553 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.048562 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.048574 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.048581 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: E0220 06:41:45.056278 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:45Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.058221 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.058245 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.058255 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.058264 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.058271 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: E0220 06:41:45.068845 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:45Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.071229 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.071252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.071260 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.071269 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.071276 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: E0220 06:41:45.078677 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:45Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:45 crc kubenswrapper[4492]: E0220 06:41:45.078781 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.121621 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.121647 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.121655 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.121666 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.121674 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.223291 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.223314 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.223323 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.223335 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.223344 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.324946 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.324971 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.324979 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.324989 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.324996 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.426068 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.426092 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.426099 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.426109 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.426116 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.527195 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.527232 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.527240 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.527255 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.527263 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.550991 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 16:06:32.203232506 +0000 UTC Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.556448 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:45 crc kubenswrapper[4492]: E0220 06:41:45.556561 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.629218 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.629242 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.629251 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.629262 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.629270 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.730782 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.730806 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.730816 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.730826 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.730833 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.832740 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.832767 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.832776 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.832786 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.832794 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.934517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.934540 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.934548 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.934557 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:45 crc kubenswrapper[4492]: I0220 06:41:45.934563 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:45Z","lastTransitionTime":"2026-02-20T06:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.036039 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.036063 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.036081 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.036090 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.036098 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.137601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.137624 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.137632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.137643 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.137650 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.239266 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.239290 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.239297 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.239307 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.239316 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.341420 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.341455 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.341463 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.341500 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.341510 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.443143 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.443164 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.443172 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.443185 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.443193 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.545406 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.545437 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.545445 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.545457 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.545465 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.551784 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 02:39:15.122185323 +0000 UTC Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.556019 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.556069 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:46 crc kubenswrapper[4492]: E0220 06:41:46.556107 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:46 crc kubenswrapper[4492]: E0220 06:41:46.556176 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.556019 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:46 crc kubenswrapper[4492]: E0220 06:41:46.556256 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.646741 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.646778 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.646787 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.646801 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.646812 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.687312 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:46 crc kubenswrapper[4492]: E0220 06:41:46.687396 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:46 crc kubenswrapper[4492]: E0220 06:41:46.687432 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:42:18.687420876 +0000 UTC m=+95.458709854 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.749037 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.749102 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.749113 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.749127 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.749136 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.850738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.850761 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.850771 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.850782 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.850791 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.952276 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.952312 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.952321 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.952334 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:46 crc kubenswrapper[4492]: I0220 06:41:46.952343 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:46Z","lastTransitionTime":"2026-02-20T06:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.054341 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.054368 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.054376 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.054387 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.054395 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.156994 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.157036 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.157045 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.157057 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.157066 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.258379 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.258409 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.258418 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.258429 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.258437 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.360272 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.360304 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.360313 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.360324 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.360332 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.461891 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.461925 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.461934 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.461947 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.461957 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.552819 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 21:29:58.720714251 +0000 UTC Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.556180 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:47 crc kubenswrapper[4492]: E0220 06:41:47.556273 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.563955 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.563983 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.563992 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.564002 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.564009 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.665321 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.665358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.665367 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.665381 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.665391 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.766756 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.766786 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.766795 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.766806 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.766816 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.868887 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.868948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.868961 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.868978 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.868990 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.970718 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.970745 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.970754 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.970766 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:47 crc kubenswrapper[4492]: I0220 06:41:47.970775 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:47Z","lastTransitionTime":"2026-02-20T06:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.072506 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.072542 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.072553 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.072567 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.072577 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.174271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.174319 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.174331 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.174343 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.174350 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.276175 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.276359 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.276368 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.276380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.276389 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.377625 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.377646 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.377655 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.377665 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.377673 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.479493 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.479522 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.479531 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.479541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.479548 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.553298 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 18:57:01.765494632 +0000 UTC Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.556589 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:48 crc kubenswrapper[4492]: E0220 06:41:48.556691 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.556591 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:48 crc kubenswrapper[4492]: E0220 06:41:48.556900 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.556990 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:48 crc kubenswrapper[4492]: E0220 06:41:48.557097 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.580972 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.580993 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.581001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.581009 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.581017 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.682167 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.682202 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.682211 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.682224 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.682232 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.783632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.783654 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.783662 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.783671 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.783678 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.857554 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/0.log" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.857645 4492 generic.go:334] "Generic (PLEG): container finished" podID="dcc11973-022a-47f9-b1e2-23e945352e7f" containerID="d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8" exitCode=1 Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.857704 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerDied","Data":"d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.858007 4492 scope.go:117] "RemoveContainer" containerID="d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.869676 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.877911 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.885141 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.885162 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.885170 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.885181 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.885189 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.888183 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.897109 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.910644 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.920384 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.928593 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.939212 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.955279 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.965689 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.974894 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.982877 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.986540 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.986591 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.986611 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.986638 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.986656 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:48Z","lastTransitionTime":"2026-02-20T06:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.991545 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:48 crc kubenswrapper[4492]: I0220 06:41:48.999907 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:48Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.010237 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.018142 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.032702 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.089380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.089420 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.089431 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.089452 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.089464 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.191099 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.191148 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.191158 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.191177 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.191189 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.292925 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.292979 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.292988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.293002 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.293011 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.395286 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.395331 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.395343 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.395363 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.395374 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.496996 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.497030 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.497038 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.497052 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.497062 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.553848 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 16:08:07.589512652 +0000 UTC Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.556154 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:49 crc kubenswrapper[4492]: E0220 06:41:49.556282 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.599314 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.599349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.599366 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.599382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.599393 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.701600 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.701635 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.701645 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.701657 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.701666 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.803439 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.803501 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.803515 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.803532 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.803546 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.861907 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/0.log" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.861965 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerStarted","Data":"d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.873968 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.887639 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.895919 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.905247 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.905798 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.905869 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.905888 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.905913 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.905932 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:49Z","lastTransitionTime":"2026-02-20T06:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.916529 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.924356 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.935063 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.943312 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.952107 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.960246 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.967902 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.977443 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.986518 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:49 crc kubenswrapper[4492]: I0220 06:41:49.995001 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:49Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.006923 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:50Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.008708 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.008732 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.008741 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.008757 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.008771 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.015170 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:50Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.022687 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:50Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.110418 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.110564 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.110647 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.110707 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.110771 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.213089 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.213143 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.213158 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.213169 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.213179 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.315500 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.315528 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.315539 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.315550 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.315559 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.416960 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.416979 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.416989 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.416998 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.417006 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.519618 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.519784 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.519869 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.519992 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.520078 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.554260 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 17:09:25.500856719 +0000 UTC Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.556549 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.556599 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:50 crc kubenswrapper[4492]: E0220 06:41:50.556662 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:50 crc kubenswrapper[4492]: E0220 06:41:50.556803 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.556869 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:50 crc kubenswrapper[4492]: E0220 06:41:50.556934 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.622178 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.622208 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.622217 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.622230 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.622240 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.723776 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.724016 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.724077 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.724135 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.724202 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.827529 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.827588 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.827604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.827626 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.827640 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.930001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.930108 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.930180 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.930251 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:50 crc kubenswrapper[4492]: I0220 06:41:50.930317 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:50Z","lastTransitionTime":"2026-02-20T06:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.032002 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.032029 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.032038 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.032048 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.032056 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.133697 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.133732 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.133775 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.133790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.133797 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.235578 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.235754 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.235826 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.235903 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.235964 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.338169 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.338216 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.338228 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.338240 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.338248 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.440035 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.440071 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.440080 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.440097 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.440108 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.541548 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.541622 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.541632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.541642 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.541650 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.554867 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 13:06:10.398721179 +0000 UTC Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.556083 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:51 crc kubenswrapper[4492]: E0220 06:41:51.556385 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.556519 4492 scope.go:117] "RemoveContainer" containerID="f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.643592 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.643629 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.643639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.643656 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.643671 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.745417 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.745462 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.745488 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.745508 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.745523 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.847095 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.847144 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.847164 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.847183 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.847202 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.869180 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/2.log" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.871023 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.871651 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.885033 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.897629 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.908346 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.924846 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.940318 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.949259 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.949294 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.949304 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.949319 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.949329 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:51Z","lastTransitionTime":"2026-02-20T06:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.951185 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.962143 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.977959 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.986199 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:51 crc kubenswrapper[4492]: I0220 06:41:51.993242 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.000264 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:51Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.010824 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.018233 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.028833 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.041560 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.050370 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.051812 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.051895 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.051907 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.051949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.051963 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.061041 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.157814 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.157852 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.157870 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.157889 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.157902 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.260151 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.260202 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.260215 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.260234 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.260248 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.362243 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.362283 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.362292 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.362308 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.362319 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.464509 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.464557 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.464570 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.464591 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.464603 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.555588 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 01:13:01.54966945 +0000 UTC Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.556800 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.556869 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.556882 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:52 crc kubenswrapper[4492]: E0220 06:41:52.556918 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:52 crc kubenswrapper[4492]: E0220 06:41:52.557018 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:52 crc kubenswrapper[4492]: E0220 06:41:52.557122 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.566315 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.566357 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.566367 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.566382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.566394 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.668491 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.668536 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.668550 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.668568 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.668581 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.770288 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.770319 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.770328 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.770341 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.770351 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.872805 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.873075 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.873086 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.873099 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.873110 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.874893 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/3.log" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.875440 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/2.log" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.877906 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" exitCode=1 Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.877938 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.877971 4492 scope.go:117] "RemoveContainer" containerID="f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.878706 4492 scope.go:117] "RemoveContainer" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" Feb 20 06:41:52 crc kubenswrapper[4492]: E0220 06:41:52.878859 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.892102 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.902697 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.911857 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.920600 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.931823 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.940699 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.951186 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.960663 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.974905 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:52Z\\\",\\\"message\\\":\\\"ng reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236198 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236197 6439 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237849 6439 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 06:41:52.237881 6439 factory.go:656] Stopping watch factory\\\\nI0220 06:41:52.236228 6439 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237903 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 06:41:52.237914 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 06:41:52.236256 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.280607 6439 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0220 06:41:52.280643 6439 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0220 06:41:52.280710 6439 ovnkube.go:599] Stopped ovnkube\\\\nI0220 06:41:52.280756 6439 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0220 06:41:52.280845 6439 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.975639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.975687 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.975704 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.975718 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.975766 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:52Z","lastTransitionTime":"2026-02-20T06:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.984924 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:52 crc kubenswrapper[4492]: I0220 06:41:52.993409 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:52Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.004502 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.012650 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.021505 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.028874 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.037659 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.045190 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.077833 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.077875 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.077885 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.077903 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.077913 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.179803 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.179843 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.179856 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.179871 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.179881 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.282289 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.282333 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.282343 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.282356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.282366 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.384199 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.384226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.384237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.384250 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.384259 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.486115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.486153 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.486164 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.486191 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.486242 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.556048 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 20:04:04.670352596 +0000 UTC Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.556325 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:53 crc kubenswrapper[4492]: E0220 06:41:53.556495 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.569035 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.581909 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.588263 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.588308 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.588323 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.588340 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.588352 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.592375 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.604361 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.613411 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.625791 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.636334 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.643273 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.651946 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.661512 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.671889 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.680359 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.690181 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.690235 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.690247 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.690266 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.690296 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.693760 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6fc70ec35a16cd67343b0871dae3c825ae25c4e9fd7cad6ea348c465875f114\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:22Z\\\",\\\"message\\\":\\\"97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804170 6023 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804172 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0220 06:41:22.804177 6023 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:52Z\\\",\\\"message\\\":\\\"ng reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236198 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236197 6439 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237849 6439 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 06:41:52.237881 6439 factory.go:656] Stopping watch factory\\\\nI0220 06:41:52.236228 6439 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237903 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 06:41:52.237914 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 06:41:52.236256 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.280607 6439 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0220 06:41:52.280643 6439 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0220 06:41:52.280710 6439 ovnkube.go:599] Stopped ovnkube\\\\nI0220 06:41:52.280756 6439 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0220 06:41:52.280845 6439 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.702984 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.712273 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.721219 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.729171 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.792665 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.792726 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.792738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.792755 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.792787 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.883883 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/3.log" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.887579 4492 scope.go:117] "RemoveContainer" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" Feb 20 06:41:53 crc kubenswrapper[4492]: E0220 06:41:53.887708 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.905601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.905626 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.905636 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.905651 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.905659 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:53Z","lastTransitionTime":"2026-02-20T06:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.908404 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.934229 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.961542 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.974615 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.986227 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:53 crc kubenswrapper[4492]: I0220 06:41:53.995426 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:53Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.004249 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.007619 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.007663 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.007674 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.007690 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.007701 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.012501 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.021735 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.031512 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.042707 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.051606 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.069286 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:52Z\\\",\\\"message\\\":\\\"ng reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236198 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236197 6439 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237849 6439 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 06:41:52.237881 6439 factory.go:656] Stopping watch factory\\\\nI0220 06:41:52.236228 6439 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237903 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 06:41:52.237914 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 06:41:52.236256 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.280607 6439 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0220 06:41:52.280643 6439 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0220 06:41:52.280710 6439 ovnkube.go:599] Stopped ovnkube\\\\nI0220 06:41:52.280756 6439 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0220 06:41:52.280845 6439 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.087950 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.096653 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.109028 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.110104 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.110201 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.110215 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.110255 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.110273 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.118558 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:54Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.212266 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.212299 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.212311 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.212327 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.212339 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.314832 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.314943 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.315006 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.315066 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.315118 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.417346 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.417382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.417391 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.417403 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.417412 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.519632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.519754 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.519906 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.520041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.520198 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.556102 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:54 crc kubenswrapper[4492]: E0220 06:41:54.556196 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.556320 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 09:05:03.102585889 +0000 UTC Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.556767 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.556825 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:54 crc kubenswrapper[4492]: E0220 06:41:54.556986 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:54 crc kubenswrapper[4492]: E0220 06:41:54.557048 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.622800 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.622835 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.622847 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.622862 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.622873 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.725836 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.725869 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.725878 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.725892 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.725901 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.828260 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.828301 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.828312 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.828326 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.828336 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.930015 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.930065 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.930079 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.930105 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:54 crc kubenswrapper[4492]: I0220 06:41:54.930120 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:54Z","lastTransitionTime":"2026-02-20T06:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.031744 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.031778 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.031788 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.031802 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.031811 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.134010 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.134063 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.134074 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.134092 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.134102 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.235811 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.235850 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.235860 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.235876 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.235887 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.258266 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.258301 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.258313 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.258326 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.258336 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: E0220 06:41:55.269054 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:55Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.275519 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.275636 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.275709 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.275775 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.275831 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: E0220 06:41:55.287452 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:55Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.291431 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.291502 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.291515 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.291538 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.291563 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: E0220 06:41:55.301659 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:55Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.305877 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.305908 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.305917 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.305929 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.305936 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: E0220 06:41:55.315749 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:55Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.318703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.318731 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.318740 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.318752 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.318759 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: E0220 06:41:55.329709 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:41:55Z is after 2025-08-24T17:21:41Z" Feb 20 06:41:55 crc kubenswrapper[4492]: E0220 06:41:55.329814 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.337833 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.337919 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.337995 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.338069 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.338130 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.439884 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.439908 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.439917 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.439929 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.439960 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.542508 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.542557 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.542569 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.542580 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.542588 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.556019 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:55 crc kubenswrapper[4492]: E0220 06:41:55.556173 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.556993 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 22:16:03.169156951 +0000 UTC Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.643986 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.644037 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.644048 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.644059 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.644067 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.746250 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.746283 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.746293 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.746305 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.746316 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.848298 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.848327 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.848338 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.848350 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.848360 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.955549 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.955579 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.955590 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.955601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:55 crc kubenswrapper[4492]: I0220 06:41:55.955612 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:55Z","lastTransitionTime":"2026-02-20T06:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.057398 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.057526 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.057601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.057670 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.057734 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.159639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.159930 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.160006 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.160076 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.160140 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.261991 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.262124 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.262201 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.262258 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.262306 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.364027 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.364050 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.364079 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.364091 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.364099 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.466540 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.466619 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.466639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.466668 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.466684 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.556191 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.556283 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.556191 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:56 crc kubenswrapper[4492]: E0220 06:41:56.556296 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:56 crc kubenswrapper[4492]: E0220 06:41:56.556504 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:56 crc kubenswrapper[4492]: E0220 06:41:56.556574 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.558343 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 04:57:21.892263294 +0000 UTC Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.569003 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.569031 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.569041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.569050 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.569058 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.670931 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.670960 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.670969 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.670983 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.671011 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.772803 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.772889 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.772905 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.772948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.773001 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.875941 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.875967 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.875978 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.875988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.875996 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.978213 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.978237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.978246 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.978258 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:56 crc kubenswrapper[4492]: I0220 06:41:56.978266 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:56Z","lastTransitionTime":"2026-02-20T06:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.080100 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.080264 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.080321 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.080382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.080492 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.182238 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.182283 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.182296 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.182315 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.182329 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.283889 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.283948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.283963 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.283983 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.284007 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.386830 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.386855 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.386863 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.386873 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.386881 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.488445 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.488468 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.488492 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.488501 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.488508 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.556932 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:57 crc kubenswrapper[4492]: E0220 06:41:57.557120 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.558509 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 05:43:23.443765806 +0000 UTC Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.590567 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.590652 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.590737 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.590828 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.590883 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.693503 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.693615 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.693668 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.693736 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.693800 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.796298 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.796356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.796367 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.796390 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.796404 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.898650 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.898689 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.898700 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.898716 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:57 crc kubenswrapper[4492]: I0220 06:41:57.898729 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:57Z","lastTransitionTime":"2026-02-20T06:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.000204 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.000233 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.000242 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.000252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.000262 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.102600 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.102628 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.102638 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.102651 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.102661 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.204512 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.204533 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.204541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.204549 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.204557 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.306796 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.306829 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.306839 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.306852 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.306862 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.408897 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.408925 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.408935 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.408948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.408957 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.510963 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.510991 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.511001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.511011 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.511019 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.556255 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:41:58 crc kubenswrapper[4492]: E0220 06:41:58.556346 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.556421 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:41:58 crc kubenswrapper[4492]: E0220 06:41:58.556515 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.556424 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:41:58 crc kubenswrapper[4492]: E0220 06:41:58.556749 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.559505 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 10:31:46.940990464 +0000 UTC Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.613179 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.613210 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.613218 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.613227 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.613235 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.715160 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.715206 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.715216 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.715226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.715233 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.818026 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.818160 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.818265 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.818362 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.818431 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.921005 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.921111 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.921171 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.921252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:58 crc kubenswrapper[4492]: I0220 06:41:58.921319 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:58Z","lastTransitionTime":"2026-02-20T06:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.023601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.023711 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.023778 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.023852 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.023933 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.125774 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.125818 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.125831 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.125846 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.125858 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.227575 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.227605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.227614 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.227625 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.227633 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.329004 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.329033 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.329041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.329053 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.329061 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.431281 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.431314 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.431323 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.431334 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.431344 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.533035 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.533218 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.533322 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.533409 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.533491 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.555993 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:41:59 crc kubenswrapper[4492]: E0220 06:41:59.556096 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.559901 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 18:07:09.318434664 +0000 UTC Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.636385 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.636523 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.636629 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.636714 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.636778 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.739322 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.739369 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.739379 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.739400 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.739411 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.841420 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.841483 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.841495 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.841513 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.841529 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.943558 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.943593 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.943604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.943616 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:41:59 crc kubenswrapper[4492]: I0220 06:41:59.943625 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:41:59Z","lastTransitionTime":"2026-02-20T06:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.045455 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.045518 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.045527 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.045541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.045552 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.147429 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.147459 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.147469 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.147765 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.147883 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.249820 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.249875 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.249887 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.249901 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.249912 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.351810 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.351839 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.351850 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.351862 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.351873 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.453156 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.453184 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.453206 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.453219 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.453229 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.555794 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.555820 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.555829 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.555839 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.555846 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.555949 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.555995 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.556046 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:00 crc kubenswrapper[4492]: E0220 06:42:00.556173 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:00 crc kubenswrapper[4492]: E0220 06:42:00.556256 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:00 crc kubenswrapper[4492]: E0220 06:42:00.556331 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.561019 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 07:17:26.959345045 +0000 UTC Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.657947 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.657988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.657996 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.658010 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.658021 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.759643 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.759679 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.759689 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.759703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.759713 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.861625 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.861679 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.861688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.861700 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.861709 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.963583 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.963615 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.963624 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.963636 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:00 crc kubenswrapper[4492]: I0220 06:42:00.963644 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:00Z","lastTransitionTime":"2026-02-20T06:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.065534 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.065587 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.065597 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.065631 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.065644 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.167463 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.167528 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.167550 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.167569 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.167592 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.269593 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.269640 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.269652 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.269666 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.269676 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.373682 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.373811 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.373882 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.373948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.373997 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.476170 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.476220 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.476230 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.476247 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.476259 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.556351 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:01 crc kubenswrapper[4492]: E0220 06:42:01.556467 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.561189 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 18:43:20.015237972 +0000 UTC Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.578577 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.578610 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.578617 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.578630 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.578642 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.680794 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.680840 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.680851 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.680869 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.680882 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.782792 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.782832 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.782845 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.782862 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.782877 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.884223 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.884261 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.884271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.884290 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.884312 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.986436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.986496 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.986508 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.986522 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:01 crc kubenswrapper[4492]: I0220 06:42:01.986531 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:01Z","lastTransitionTime":"2026-02-20T06:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.088375 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.088413 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.088424 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.088443 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.088457 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.190166 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.190220 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.190233 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.190249 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.190263 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.291604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.291632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.291640 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.291654 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.291664 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.393277 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.393324 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.393332 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.393345 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.393355 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.495116 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.495158 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.495166 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.495181 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.495206 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.556228 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:02 crc kubenswrapper[4492]: E0220 06:42:02.556339 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.556369 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.556407 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:02 crc kubenswrapper[4492]: E0220 06:42:02.556460 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:02 crc kubenswrapper[4492]: E0220 06:42:02.556516 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.562323 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 04:54:42.58428708 +0000 UTC Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.569829 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.597881 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.597927 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.597939 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.597953 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.597970 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.700551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.700605 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.700616 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.700630 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.700656 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.802685 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.802712 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.802720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.802732 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.802766 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.905040 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.905068 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.905076 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.905087 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:02 crc kubenswrapper[4492]: I0220 06:42:02.905096 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:02Z","lastTransitionTime":"2026-02-20T06:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.007821 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.007841 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.007849 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.007858 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.007865 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.109076 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.109102 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.109110 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.109123 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.109137 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.211188 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.211236 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.211245 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.211254 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.211262 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.313151 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.313191 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.313211 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.313226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.313235 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.415134 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.415166 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.415175 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.415185 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.415193 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.517208 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.517269 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.517279 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.517300 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.517315 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.556662 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:03 crc kubenswrapper[4492]: E0220 06:42:03.556825 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.562824 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 16:38:55.097993533 +0000 UTC Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.572436 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.581864 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.597258 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:52Z\\\",\\\"message\\\":\\\"ng reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236198 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236197 6439 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237849 6439 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 06:41:52.237881 6439 factory.go:656] Stopping watch factory\\\\nI0220 06:41:52.236228 6439 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237903 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 06:41:52.237914 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 06:41:52.236256 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.280607 6439 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0220 06:41:52.280643 6439 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0220 06:41:52.280710 6439 ovnkube.go:599] Stopped ovnkube\\\\nI0220 06:41:52.280756 6439 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0220 06:41:52.280845 6439 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.607220 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.617020 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.618872 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.618902 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.618912 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.618925 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.618941 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.629154 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.639808 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.647269 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.660552 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02fb15fd-3a70-4ac3-bd10-844911a2d613\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad97b83da8467370c50107a9d345df8e4f74b9d0b4396d3ee8899696be74272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa8901bcd48ad3e0f9ea540e2a05e1177ec64c608504a8f91e02f2a38950d44e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c82f1f4abdf61156fe0997903135d59828c1913e7d5dc56ccd863c0af9854766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693404fcaf966a707f4d5ee5f72466873e33188f019eb07a819e52d774b489f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a46e82a3a9c08ce424bb59b1720fe552bb5bddc3ae515a52eab7b71cfd9b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda42d8dd09c47cbf9ee70cfa621c6eb7d03ec9277d91257c14322e73a0d5d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda42d8dd09c47cbf9ee70cfa621c6eb7d03ec9277d91257c14322e73a0d5d1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08368f8dfa1ff094e57f746f499d5aa6f0fd9a2e1e94ef7339cb5f47d99bd96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08368f8dfa1ff094e57f746f499d5aa6f0fd9a2e1e94ef7339cb5f47d99bd96e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4af8fd874cb5c56d88a7855415b20d5ece4d7ab723f6d779019be494a6c2451c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4af8fd874cb5c56d88a7855415b20d5ece4d7ab723f6d779019be494a6c2451c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.669164 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.677810 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.685163 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.693249 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.700621 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.710340 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.718085 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.720429 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.720458 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.720499 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.720515 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.720528 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.726308 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.736152 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:03Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.822766 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.822801 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.822811 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.822825 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.822836 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.924419 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.924450 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.924464 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.924496 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:03 crc kubenswrapper[4492]: I0220 06:42:03.924508 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:03Z","lastTransitionTime":"2026-02-20T06:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.026033 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.026073 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.026083 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.026098 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.026109 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.128188 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.128227 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.128237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.128248 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.128255 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.230461 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.230674 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.230688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.230705 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.230716 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.332777 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.332833 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.332845 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.332859 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.332869 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.434491 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.434517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.434524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.434534 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.434540 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.443704 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.443844 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.443828661 +0000 UTC m=+145.215117639 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.541092 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.541214 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.541275 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.541331 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.541385 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.544785 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.544841 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.544865 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.544888 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.544962 4492 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.544991 4492 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545009 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545128 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545146 4492 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545009 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545014 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.545001583 +0000 UTC m=+145.316290562 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545279 4492 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545331 4492 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545311 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.545294627 +0000 UTC m=+145.316583605 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545405 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.545370981 +0000 UTC m=+145.316659959 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.545434 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.545419081 +0000 UTC m=+145.316708059 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.555962 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.556153 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.556027 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.556373 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.555981 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:04 crc kubenswrapper[4492]: E0220 06:42:04.556542 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.563096 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 07:58:55.033017874 +0000 UTC Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.642961 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.642995 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.643004 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.643017 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.643027 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.745524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.745549 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.745560 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.745572 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.745583 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.847532 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.847577 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.847589 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.847611 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.847629 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.949415 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.949451 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.949459 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.949494 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:04 crc kubenswrapper[4492]: I0220 06:42:04.949507 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:04Z","lastTransitionTime":"2026-02-20T06:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.051749 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.051773 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.051784 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.051796 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.051806 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.153720 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.153762 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.153770 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.153790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.153800 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.256053 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.256090 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.256102 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.256120 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.256133 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.357840 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.357863 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.357871 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.357882 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.357910 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.459314 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.459338 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.459346 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.459357 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.459365 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.484067 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.484090 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.484099 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.484109 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.484117 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: E0220 06:42:05.496345 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.499564 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.499615 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.499641 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.499651 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.499661 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: E0220 06:42:05.509641 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.512700 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.512726 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.512734 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.512743 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.512750 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: E0220 06:42:05.521731 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.524047 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.524084 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.524096 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.524114 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.524125 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: E0220 06:42:05.534662 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.537210 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.537242 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.537254 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.537269 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.537280 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: E0220 06:42:05.546901 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:05Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:05 crc kubenswrapper[4492]: E0220 06:42:05.547118 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.556270 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:05 crc kubenswrapper[4492]: E0220 06:42:05.556528 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.560611 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.560643 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.560655 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.560669 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.560680 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.563888 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 07:23:23.965015209 +0000 UTC Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.565505 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.662207 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.662239 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.662248 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.662261 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.662271 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.764349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.764389 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.764401 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.764422 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.764433 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.865999 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.866108 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.866172 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.866248 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.866307 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.968416 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.968455 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.968466 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.968510 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:05 crc kubenswrapper[4492]: I0220 06:42:05.968523 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:05Z","lastTransitionTime":"2026-02-20T06:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.070551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.070606 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.070618 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.070656 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.070671 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.173158 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.173433 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.173524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.173590 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.173656 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.277438 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.277507 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.277522 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.277544 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.277561 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.380060 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.380104 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.380115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.380128 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.380140 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.482091 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.482120 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.482137 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.482151 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.482162 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.556825 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.556927 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.557065 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:06 crc kubenswrapper[4492]: E0220 06:42:06.557065 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:06 crc kubenswrapper[4492]: E0220 06:42:06.557234 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:06 crc kubenswrapper[4492]: E0220 06:42:06.557401 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.564256 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 19:08:07.030006492 +0000 UTC Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.583823 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.583849 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.583858 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.583870 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.583883 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.685841 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.685868 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.685877 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.685888 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.685898 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.788304 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.788344 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.788356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.788373 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.788383 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.889904 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.889937 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.889949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.889963 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.889975 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.992310 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.992347 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.992355 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.992367 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:06 crc kubenswrapper[4492]: I0220 06:42:06.992377 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:06Z","lastTransitionTime":"2026-02-20T06:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.094196 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.094236 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.094244 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.094254 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.094265 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.196176 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.196197 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.196343 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.196354 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.196362 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.297654 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.297684 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.297694 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.297703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.297712 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.401258 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.401286 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.401295 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.401307 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.401317 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.504738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.504770 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.504791 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.504805 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.504817 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.556326 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:07 crc kubenswrapper[4492]: E0220 06:42:07.556441 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.557118 4492 scope.go:117] "RemoveContainer" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" Feb 20 06:42:07 crc kubenswrapper[4492]: E0220 06:42:07.557324 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.565025 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 23:35:39.478218736 +0000 UTC Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.606651 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.606674 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.606683 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.606692 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.606700 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.708591 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.708632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.708644 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.708662 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.708677 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.810734 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.810761 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.810769 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.810782 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.810790 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.912636 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.912674 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.912687 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.912703 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:07 crc kubenswrapper[4492]: I0220 06:42:07.912717 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:07Z","lastTransitionTime":"2026-02-20T06:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.014890 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.014926 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.014936 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.014949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.014964 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.116931 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.116972 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.116986 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.117001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.117013 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.218723 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.218745 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.218753 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.218767 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.218775 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.320036 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.320152 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.320226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.320282 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.320329 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.421764 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.421816 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.421825 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.421838 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.421848 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.523662 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.523685 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.523695 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.523706 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.523713 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.556100 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.556166 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:08 crc kubenswrapper[4492]: E0220 06:42:08.556262 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.556375 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:08 crc kubenswrapper[4492]: E0220 06:42:08.556515 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:08 crc kubenswrapper[4492]: E0220 06:42:08.556665 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.565573 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 11:59:13.027536066 +0000 UTC Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.625099 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.625126 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.625137 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.625150 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.625159 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.726970 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.726997 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.727006 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.727019 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.727030 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.829093 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.829119 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.829129 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.829139 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.829148 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.930936 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.930989 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.931000 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.931017 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:08 crc kubenswrapper[4492]: I0220 06:42:08.931029 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:08Z","lastTransitionTime":"2026-02-20T06:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.033277 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.033323 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.033335 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.033355 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.033369 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.135421 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.135458 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.135467 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.135496 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.135527 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.237161 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.237319 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.237435 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.237551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.237615 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.341395 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.341439 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.341452 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.341471 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.341504 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.443208 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.443261 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.443276 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.443307 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.443323 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.545265 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.545299 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.545311 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.545328 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.545338 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.556653 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:09 crc kubenswrapper[4492]: E0220 06:42:09.556777 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.566564 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 10:30:43.406280453 +0000 UTC Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.647184 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.647332 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.647402 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.647454 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.647535 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.749137 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.749157 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.749166 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.749180 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.749190 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.851157 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.851273 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.851356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.851424 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.851509 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.953313 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.953350 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.953358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.953373 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:09 crc kubenswrapper[4492]: I0220 06:42:09.953381 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:09Z","lastTransitionTime":"2026-02-20T06:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.055680 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.055727 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.055740 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.055759 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.055770 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.158908 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.158963 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.158974 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.158988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.159001 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.260649 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.260700 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.260712 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.260729 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.260740 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.362349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.362399 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.362411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.362424 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.362442 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.464622 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.464644 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.464655 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.464691 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.464701 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.556176 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.556181 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.556181 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:10 crc kubenswrapper[4492]: E0220 06:42:10.556357 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:10 crc kubenswrapper[4492]: E0220 06:42:10.556450 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:10 crc kubenswrapper[4492]: E0220 06:42:10.556541 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.566329 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.566363 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.566375 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.566387 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.566398 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.566707 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 09:57:26.923598147 +0000 UTC Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.668075 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.668126 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.668137 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.668152 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.668163 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.770028 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.770063 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.770073 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.770084 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.770092 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.871468 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.871518 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.871527 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.871540 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.871552 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.973885 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.973921 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.973930 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.973943 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:10 crc kubenswrapper[4492]: I0220 06:42:10.973953 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:10Z","lastTransitionTime":"2026-02-20T06:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.075883 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.075926 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.075935 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.075948 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.075957 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.177824 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.177869 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.177880 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.177894 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.177906 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.279413 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.279447 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.279458 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.279495 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.279510 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.380904 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.380933 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.380942 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.380955 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.380966 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.483318 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.483349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.483361 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.483372 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.483380 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.556617 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:11 crc kubenswrapper[4492]: E0220 06:42:11.556815 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.567172 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 23:55:48.062731701 +0000 UTC Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.585514 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.585546 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.585556 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.585568 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.585579 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.687614 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.687650 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.687659 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.687671 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.687682 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.789664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.789704 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.789712 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.789727 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.789739 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.891790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.891847 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.891858 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.891875 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.891898 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.993583 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.993616 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.993625 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.993639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:11 crc kubenswrapper[4492]: I0220 06:42:11.993648 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:11Z","lastTransitionTime":"2026-02-20T06:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.095702 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.095736 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.095744 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.095756 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.095764 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.197496 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.197524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.197532 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.197544 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.197552 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.299230 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.299269 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.299277 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.299292 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.299300 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.401135 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.401173 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.401183 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.401196 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.401204 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.503490 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.503537 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.503546 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.503562 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.503581 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.556031 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.556065 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.556034 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:12 crc kubenswrapper[4492]: E0220 06:42:12.556151 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:12 crc kubenswrapper[4492]: E0220 06:42:12.556213 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:12 crc kubenswrapper[4492]: E0220 06:42:12.556292 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.567286 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 08:55:38.734148859 +0000 UTC Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.605946 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.605987 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.605996 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.606012 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.606023 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.707710 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.707741 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.707750 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.707763 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.707773 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.810190 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.810239 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.810248 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.810259 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.810269 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.912671 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.912710 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.912719 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.912736 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:12 crc kubenswrapper[4492]: I0220 06:42:12.912746 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:12Z","lastTransitionTime":"2026-02-20T06:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.015027 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.015115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.015128 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.015154 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.015186 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.118114 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.118153 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.118162 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.118177 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.118192 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.220101 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.220147 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.220159 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.220174 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.220185 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.322558 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.322590 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.322601 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.322616 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.322628 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.425500 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.425561 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.425573 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.425594 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.425610 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.527944 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.527976 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.527987 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.528003 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.528013 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.556627 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:13 crc kubenswrapper[4492]: E0220 06:42:13.556851 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.568448 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:21:00.550652996 +0000 UTC Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.578332 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b4455c-a841-4cbd-931d-c60d3f2ec625\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5620ee6a276c2f2363f504ebcf2d8a5c18e16ca5ed6c6ad9142d8103299add6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://62a040f3596b526ae3f68da7c51075ac64864afe8b2972309272c5e000bc52e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://723c707754c6670c238f3d66bfbdac477e62d030b30dd35da39194ec230f7c93\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.587551 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"103071c4-e0aa-4f9b-bf7d-1151f7c69931\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1917ab54cbd788634304d3d873aa19de6f192f33a14914cb79b7ed4b75359cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30c66318f267832d58e50e1fb600d6e931bce4ab40190240bbcf08e2b805776e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2dc4fcd6ad07ba2d87160d48ad725a325dba0dcb193ad5a854769dbf96614c73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c19f46a27cd9b102c40b21704294e003b42be33236337fd9e87597904d905bd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.596773 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eaaad92d12717837f56220a4458fb058778d651b8c13fb7dab19c12e8e0fbc7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.605178 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.620405 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28d6c67f-c4cd-4692-b490-b2b884c72db6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:52Z\\\",\\\"message\\\":\\\"ng reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236198 6439 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.236197 6439 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237849 6439 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0220 06:41:52.237881 6439 factory.go:656] Stopping watch factory\\\\nI0220 06:41:52.236228 6439 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.237903 6439 handler.go:208] Removed *v1.Node event handler 2\\\\nI0220 06:41:52.237914 6439 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0220 06:41:52.236256 6439 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0220 06:41:52.280607 6439 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0220 06:41:52.280643 6439 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0220 06:41:52.280710 6439 ovnkube.go:599] Stopped ovnkube\\\\nI0220 06:41:52.280756 6439 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0220 06:41:52.280845 6439 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bhn54\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f68mj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.629180 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.629226 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.629237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.629252 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.629261 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.635360 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02fb15fd-3a70-4ac3-bd10-844911a2d613\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad97b83da8467370c50107a9d345df8e4f74b9d0b4396d3ee8899696be74272\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa8901bcd48ad3e0f9ea540e2a05e1177ec64c608504a8f91e02f2a38950d44e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c82f1f4abdf61156fe0997903135d59828c1913e7d5dc56ccd863c0af9854766\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://693404fcaf966a707f4d5ee5f72466873e33188f019eb07a819e52d774b489f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a46e82a3a9c08ce424bb59b1720fe552bb5bddc3ae515a52eab7b71cfd9b49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda42d8dd09c47cbf9ee70cfa621c6eb7d03ec9277d91257c14322e73a0d5d1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eda42d8dd09c47cbf9ee70cfa621c6eb7d03ec9277d91257c14322e73a0d5d1f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://08368f8dfa1ff094e57f746f499d5aa6f0fd9a2e1e94ef7339cb5f47d99bd96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08368f8dfa1ff094e57f746f499d5aa6f0fd9a2e1e94ef7339cb5f47d99bd96e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4af8fd874cb5c56d88a7855415b20d5ece4d7ab723f6d779019be494a6c2451c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4af8fd874cb5c56d88a7855415b20d5ece4d7ab723f6d779019be494a6c2451c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.644938 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.653527 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.664132 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65a8a150-0190-4595-8538-2c4c5875ba54\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://111b18fe7b5e407c292233c2940f0e8341d5f64d158a7676c3e0475228f7ed7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7a71a69fd9105b85be3045eb9d07adfee002d9476f481d903c88787a628e512\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9c1b23fb00719db9e27b752153ddf027d327c6415631cb7ef58f9300ed2263\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://83dc201f16361de4ee9e353a68e5f86bd2037ae3808aec5823a4627229fd58e0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f890dad18c9dfcae4074509d4c1018a78dec06c8382c16827454ac22a814b422\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd8001788d37a48ae11b653a1dca2011261db108e1455e08234115a93920e691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a843a67f24a7b730ba49045244515740245b386a832cfe572ea1e09ba34ae7b4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qgw7t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-pl2t8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.671583 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0825f068-2c78-44a0-8f24-679e4521f069\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4dn6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p9vbm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.680815 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c91c2a5f5307fae12f472e7f81d891503d6efb21832c32627f382013d19609\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.689168 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-xbm4p" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8b630176-8730-4b5a-bf14-46bf2bc2862d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://418b5486c1475f742b4e72b3b3e821e5f8ab609195fb55bfe7fc9d39b636770b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cbqjh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-xbm4p\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.700960 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2rr9j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcc11973-022a-47f9-b1e2-23e945352e7f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-20T06:41:48Z\\\",\\\"message\\\":\\\"2026-02-20T06:41:02+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d\\\\n2026-02-20T06:41:02+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1a6f919c-c27e-4889-a2de-a63a4c8a302d to /host/opt/cni/bin/\\\\n2026-02-20T06:41:03Z [verbose] multus-daemon started\\\\n2026-02-20T06:41:03Z [verbose] Readiness Indicator file check\\\\n2026-02-20T06:41:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:41:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8khkx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2rr9j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.708665 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-p8vnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2862f4a6-65d4-4621-acb8-78b9c13906dc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c770796dce8dbce9350ed23c6ffa1202126ea2f1f9abdc6ca996e4a8d3623e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2gtfn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:03Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-p8vnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.717386 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9891935-2eec-49bc-a744-2c5901a88d07\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a201f8dd6fd229ffc0d770b6340de35a4e97601ce09a8f5d3ac92bb67c574aa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5682c4869082fa1acab86f75e69fd5d1ef91aa65f3f95ac466cec50693a79abf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5682c4869082fa1acab86f75e69fd5d1ef91aa65f3f95ac466cec50693a79abf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.731237 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.731280 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.731293 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.731309 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.731320 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.731536 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33d8f5bf-8c00-4260-9e5b-e188570d4715\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:40:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-20T06:41:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0220 06:41:00.352016 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0220 06:41:00.352233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0220 06:41:00.353657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1670003042/tls.crt::/tmp/serving-cert-1670003042/tls.key\\\\\\\"\\\\nI0220 06:41:00.667668 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0220 06:41:00.688132 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0220 06:41:00.688168 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0220 06:41:00.688199 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0220 06:41:00.688204 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0220 06:41:00.709679 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0220 06:41:00.709701 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709706 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0220 06:41:00.709710 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0220 06:41:00.709713 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0220 06:41:00.709715 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0220 06:41:00.709718 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0220 06:41:00.709962 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0220 06:41:00.714997 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:40:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-20T06:40:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-20T06:40:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:40:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.741815 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae511fd64c687593136ce8d8cb20bf550faef71e00df9ab76fe93d0175db75ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://539bdd0bd87e5e9c77dd99f66705925f91cd98572c90706afd49720c5f37830d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.751624 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8755354-8581-4b5d-9b7e-6b7c53707f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d00f70d7bec3ce11dce43ea32ac0f0aa0b80110f8bc88ce1425ca57557b43f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nqc6l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zrlsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.760147 4492 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"921784a1-2034-4ca6-b954-af969bfefd1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-20T06:41:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcfa0a7ea251e9c57addc76f7538532014dea715d4fb4a7fe69ecd10dabcce07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58d70b3a60ec7de68acdde3e00d2c2a301472e33592574e4fc0693ca8cfe32d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-20T06:41:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q64zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-20T06:41:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-c6ffl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:13Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.833762 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.833817 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.833829 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.833842 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.833851 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.935338 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.935371 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.935381 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.935395 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:13 crc kubenswrapper[4492]: I0220 06:42:13.935403 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:13Z","lastTransitionTime":"2026-02-20T06:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.037295 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.037318 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.037329 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.037343 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.037354 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.139326 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.139349 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.139358 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.139369 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.139377 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.241598 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.241623 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.241633 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.241645 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.241653 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.343484 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.343539 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.343551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.343561 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.343570 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.445776 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.445818 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.445828 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.445838 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.445846 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.547639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.547664 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.547671 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.547680 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.547687 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.555965 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:14 crc kubenswrapper[4492]: E0220 06:42:14.556062 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.556219 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:14 crc kubenswrapper[4492]: E0220 06:42:14.556293 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.556603 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:14 crc kubenswrapper[4492]: E0220 06:42:14.556820 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.569212 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 04:34:52.679389642 +0000 UTC Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.649441 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.649491 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.649503 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.649517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.649526 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.751205 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.751246 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.751256 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.751266 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.751277 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.852958 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.852993 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.853002 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.853020 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.853033 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.955263 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.955302 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.955311 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.955330 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:14 crc kubenswrapper[4492]: I0220 06:42:14.955346 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:14Z","lastTransitionTime":"2026-02-20T06:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.057709 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.057739 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.057749 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.057767 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.057782 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.159378 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.159407 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.159434 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.159448 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.159458 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.261356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.261382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.261413 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.261425 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.261434 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.363643 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.363772 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.363836 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.363904 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.363960 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.465713 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.465763 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.465774 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.465786 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.465794 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.556617 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:15 crc kubenswrapper[4492]: E0220 06:42:15.556915 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.568545 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.568592 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.568603 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.568623 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.568634 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.569575 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 02:21:13.473484248 +0000 UTC Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.624688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.624735 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.624747 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.624769 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.624784 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: E0220 06:42:15.635429 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.638524 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.638567 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.638583 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.638598 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.638608 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: E0220 06:42:15.657065 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.660134 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.660173 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.660185 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.660199 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.660213 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: E0220 06:42:15.669878 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.673248 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.673289 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.673302 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.673316 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.673324 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: E0220 06:42:15.682978 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.686105 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.686153 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.686164 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.686178 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.686188 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: E0220 06:42:15.694965 4492 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148072Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608872Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-20T06:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cafa21b1-6432-4073-bf29-13d2dbbad80c\\\",\\\"systemUUID\\\":\\\"b34febcf-795b-4446-bb24-1edd1f924aef\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-20T06:42:15Z is after 2025-08-24T17:21:41Z" Feb 20 06:42:15 crc kubenswrapper[4492]: E0220 06:42:15.695107 4492 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.696362 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.696399 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.696411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.696430 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.696441 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.798277 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.798306 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.798317 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.798330 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.798340 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.899818 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.899857 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.899868 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.899882 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:15 crc kubenswrapper[4492]: I0220 06:42:15.899894 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:15Z","lastTransitionTime":"2026-02-20T06:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.002031 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.002055 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.002063 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.002073 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.002081 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.103901 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.103949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.103958 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.103971 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.103979 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.206847 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.206876 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.206886 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.206897 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.206906 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.308271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.308541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.308614 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.308711 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.308796 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.410566 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.410596 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.410609 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.410624 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.410633 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.512506 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.512546 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.512557 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.512569 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.512580 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.556020 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.556083 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.556080 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:16 crc kubenswrapper[4492]: E0220 06:42:16.556187 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:16 crc kubenswrapper[4492]: E0220 06:42:16.556359 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:16 crc kubenswrapper[4492]: E0220 06:42:16.556406 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.570239 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 03:53:56.148129686 +0000 UTC Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.615411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.615463 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.615507 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.615523 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.615532 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.717581 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.717614 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.717623 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.717639 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.717649 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.818999 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.819021 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.819032 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.819045 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.819055 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.920887 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.920940 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.920951 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.920962 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:16 crc kubenswrapper[4492]: I0220 06:42:16.920973 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:16Z","lastTransitionTime":"2026-02-20T06:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.023280 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.023313 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.023324 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.023359 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.023369 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.125514 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.125561 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.125576 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.125595 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.125605 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.227945 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.228001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.228015 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.228026 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.228038 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.329904 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.330030 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.330096 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.330178 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.330267 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.432244 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.432400 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.432498 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.432560 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.432628 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.534660 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.534711 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.534726 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.534750 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.534767 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.555966 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:17 crc kubenswrapper[4492]: E0220 06:42:17.556245 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.570923 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 09:13:55.311810711 +0000 UTC Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.636688 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.636742 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.636755 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.636771 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.636784 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.739128 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.739171 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.739181 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.739194 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.739202 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.840928 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.840964 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.840975 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.840991 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.841001 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.942970 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.943007 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.943018 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.943034 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:17 crc kubenswrapper[4492]: I0220 06:42:17.943048 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:17Z","lastTransitionTime":"2026-02-20T06:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.044535 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.044644 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.044726 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.044793 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.044844 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.146445 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.146506 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.146517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.146529 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.146539 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.248870 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.248998 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.249080 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.249144 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.249201 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.350842 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.350968 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.351029 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.351092 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.351148 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.453139 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.453271 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.453336 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.453498 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.453566 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.555982 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.555987 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.556054 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.556287 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.556319 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.556330 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.556344 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.556353 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: E0220 06:42:18.556634 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.556800 4492 scope.go:117] "RemoveContainer" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" Feb 20 06:42:18 crc kubenswrapper[4492]: E0220 06:42:18.556806 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:18 crc kubenswrapper[4492]: E0220 06:42:18.556894 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:18 crc kubenswrapper[4492]: E0220 06:42:18.556977 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.571901 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 07:25:41.541808284 +0000 UTC Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.657903 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.657946 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.657957 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.657972 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.657983 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.760054 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.760093 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.760106 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.760121 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.760132 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.777457 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:18 crc kubenswrapper[4492]: E0220 06:42:18.777610 4492 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:42:18 crc kubenswrapper[4492]: E0220 06:42:18.777692 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs podName:0825f068-2c78-44a0-8f24-679e4521f069 nodeName:}" failed. No retries permitted until 2026-02-20 06:43:22.77767146 +0000 UTC m=+159.548960428 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs") pod "network-metrics-daemon-p9vbm" (UID: "0825f068-2c78-44a0-8f24-679e4521f069") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.862178 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.862303 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.862382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.862459 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.862546 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.964738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.964769 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.964777 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.964790 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:18 crc kubenswrapper[4492]: I0220 06:42:18.964799 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:18Z","lastTransitionTime":"2026-02-20T06:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.066562 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.066602 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.066613 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.066630 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.066642 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.168429 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.168567 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.168656 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.168734 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.168794 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.271118 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.271245 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.271307 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.271380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.271455 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.373293 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.373334 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.373348 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.373363 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.373375 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.474516 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.474660 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.474758 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.474910 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.475017 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.556307 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:19 crc kubenswrapper[4492]: E0220 06:42:19.556553 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.572162 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 17:17:20.334261653 +0000 UTC Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.576332 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.576446 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.576551 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.576632 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.576703 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.678987 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.679128 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.679214 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.679305 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.679399 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.783030 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.783065 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.783075 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.783089 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.783131 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.884983 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.885027 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.885041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.885062 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.885074 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.986850 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.986890 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.986925 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.986937 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:19 crc kubenswrapper[4492]: I0220 06:42:19.986946 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:19Z","lastTransitionTime":"2026-02-20T06:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.089130 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.089178 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.089191 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.089205 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.089215 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.190843 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.190883 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.190893 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.190910 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.190922 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.292848 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.292885 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.292893 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.292907 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.292917 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.395208 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.395263 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.395276 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.395291 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.395305 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.497059 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.497093 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.497105 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.497117 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.497125 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.556148 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.556175 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.556207 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:20 crc kubenswrapper[4492]: E0220 06:42:20.556280 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:20 crc kubenswrapper[4492]: E0220 06:42:20.556377 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:20 crc kubenswrapper[4492]: E0220 06:42:20.556456 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.572466 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 00:32:45.451974368 +0000 UTC Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.599687 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.599743 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.599755 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.599771 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.599785 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.702361 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.702408 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.702420 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.702436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.702451 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.804411 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.804454 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.804463 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.804497 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.804508 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.906091 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.906122 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.906132 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.906146 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:20 crc kubenswrapper[4492]: I0220 06:42:20.906156 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:20Z","lastTransitionTime":"2026-02-20T06:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.007907 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.007936 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.007946 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.007958 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.007968 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.109710 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.109762 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.109779 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.109789 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.109800 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.211391 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.211436 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.211446 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.211460 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.211493 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.313412 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.313465 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.313499 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.313522 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.313537 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.415181 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.415231 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.415257 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.415278 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.415297 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.517734 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.517786 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.517798 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.517818 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.517831 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.555945 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:21 crc kubenswrapper[4492]: E0220 06:42:21.556055 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.573373 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 15:31:34.787733167 +0000 UTC Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.620109 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.620138 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.620147 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.620163 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.620174 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.722074 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.722094 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.722103 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.722115 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.722126 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.824431 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.824461 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.824488 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.824520 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.824530 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.926701 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.926819 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.926896 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.926974 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:21 crc kubenswrapper[4492]: I0220 06:42:21.927036 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:21Z","lastTransitionTime":"2026-02-20T06:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.029007 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.029049 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.029062 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.029075 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.029084 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.130907 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.130961 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.130972 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.130989 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.131003 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.232288 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.232322 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.232333 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.232346 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.232356 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.334539 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.334585 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.334599 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.334613 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.334621 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.436281 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.436303 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.436315 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.436346 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.436355 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.538766 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.538889 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.538984 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.539062 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.539140 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.556002 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.556059 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.556219 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:22 crc kubenswrapper[4492]: E0220 06:42:22.556376 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:22 crc kubenswrapper[4492]: E0220 06:42:22.556689 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:22 crc kubenswrapper[4492]: E0220 06:42:22.556801 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.573654 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 06:44:36.280486266 +0000 UTC Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.641702 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.642132 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.642206 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.642291 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.642356 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.744533 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.744577 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.744588 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.744604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.744614 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.846530 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.846560 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.846590 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.846604 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.846611 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.948779 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.948809 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.948819 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.948833 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:22 crc kubenswrapper[4492]: I0220 06:42:22.948844 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:22Z","lastTransitionTime":"2026-02-20T06:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.050727 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.050761 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.050772 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.050785 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.050794 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.153287 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.153337 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.153365 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.153382 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.153392 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.255041 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.255120 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.255131 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.255146 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.255157 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.357381 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.357416 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.357426 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.357439 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.357447 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.458738 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.458776 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.458786 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.458802 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.458811 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.556550 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:23 crc kubenswrapper[4492]: E0220 06:42:23.556664 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.559805 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.559824 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.559835 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.559846 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.559855 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.574596 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:42:20.322272243 +0000 UTC Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.608109 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.608092586 podStartE2EDuration="1m19.608092586s" podCreationTimestamp="2026-02-20 06:41:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.607583526 +0000 UTC m=+100.378872514" watchObservedRunningTime="2026-02-20 06:42:23.608092586 +0000 UTC m=+100.379381554" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.617048 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=46.61702628 podStartE2EDuration="46.61702628s" podCreationTimestamp="2026-02-20 06:41:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.616383187 +0000 UTC m=+100.387672165" watchObservedRunningTime="2026-02-20 06:42:23.61702628 +0000 UTC m=+100.388315258" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.648058 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pl2t8" podStartSLOduration=82.648038224 podStartE2EDuration="1m22.648038224s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.640562481 +0000 UTC m=+100.411851459" watchObservedRunningTime="2026-02-20 06:42:23.648038224 +0000 UTC m=+100.419327202" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.662261 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.662491 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.662588 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.662675 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.662748 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.666025 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=21.666012933 podStartE2EDuration="21.666012933s" podCreationTimestamp="2026-02-20 06:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.665451404 +0000 UTC m=+100.436740382" watchObservedRunningTime="2026-02-20 06:42:23.666012933 +0000 UTC m=+100.437301911" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.683334 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2rr9j" podStartSLOduration=82.683319452 podStartE2EDuration="1m22.683319452s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.683006843 +0000 UTC m=+100.454295821" watchObservedRunningTime="2026-02-20 06:42:23.683319452 +0000 UTC m=+100.454608430" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.690575 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-p8vnd" podStartSLOduration=83.690564641 podStartE2EDuration="1m23.690564641s" podCreationTimestamp="2026-02-20 06:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.690365855 +0000 UTC m=+100.461654833" watchObservedRunningTime="2026-02-20 06:42:23.690564641 +0000 UTC m=+100.461853618" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.715411 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-xbm4p" podStartSLOduration=83.715397548 podStartE2EDuration="1m23.715397548s" podCreationTimestamp="2026-02-20 06:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.706961032 +0000 UTC m=+100.478250011" watchObservedRunningTime="2026-02-20 06:42:23.715397548 +0000 UTC m=+100.486686526" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.722993 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podStartSLOduration=82.722983599 podStartE2EDuration="1m22.722983599s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.72260262 +0000 UTC m=+100.493891598" watchObservedRunningTime="2026-02-20 06:42:23.722983599 +0000 UTC m=+100.494272576" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.739581 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=18.739563297 podStartE2EDuration="18.739563297s" podCreationTimestamp="2026-02-20 06:42:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.739303647 +0000 UTC m=+100.510592625" watchObservedRunningTime="2026-02-20 06:42:23.739563297 +0000 UTC m=+100.510852275" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.740232 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-c6ffl" podStartSLOduration=82.740226297 podStartE2EDuration="1m22.740226297s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.733099383 +0000 UTC m=+100.504388361" watchObservedRunningTime="2026-02-20 06:42:23.740226297 +0000 UTC m=+100.511515276" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.764991 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.765046 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.765056 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.765072 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.765084 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.868037 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.868076 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.868104 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.868121 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.868130 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.970499 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.970530 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.970541 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.970555 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:23 crc kubenswrapper[4492]: I0220 06:42:23.970563 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:23Z","lastTransitionTime":"2026-02-20T06:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.071988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.072022 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.072030 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.072043 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.072052 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.173973 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.174130 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.174196 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.174277 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.174340 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.276240 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.276300 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.276310 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.276323 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.276331 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.377907 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.377940 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.377949 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.377961 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.377972 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.479988 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.480016 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.480024 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.480035 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.480043 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.556875 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.556907 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.556932 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:24 crc kubenswrapper[4492]: E0220 06:42:24.557022 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:24 crc kubenswrapper[4492]: E0220 06:42:24.557106 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:24 crc kubenswrapper[4492]: E0220 06:42:24.557162 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.575272 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 09:54:41.214437515 +0000 UTC Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.581539 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.581564 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.581575 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.581586 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.581595 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.683545 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.683574 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.683583 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.683592 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.683600 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.785721 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.785744 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.785752 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.785763 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.785770 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.887675 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.887701 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.887709 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.887724 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.887732 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.988942 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.988964 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.988973 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.988984 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:24 crc kubenswrapper[4492]: I0220 06:42:24.988992 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:24Z","lastTransitionTime":"2026-02-20T06:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.090313 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.090337 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.090344 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.090353 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.090359 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.192427 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.192452 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.192460 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.192487 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.192497 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.294623 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.294652 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.294667 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.294677 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.294686 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.396426 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.396464 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.396489 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.396517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.396528 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.498517 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.498789 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.498861 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.498925 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.498990 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.556159 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:25 crc kubenswrapper[4492]: E0220 06:42:25.556495 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.575863 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 05:33:19.051102408 +0000 UTC Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.600980 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.601001 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.601012 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.601023 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.601031 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.702514 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.702531 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.702538 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.702547 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.702554 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.802912 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.802946 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.802955 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.802968 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.802975 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.813356 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.813380 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.813386 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.813395 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.813402 4492 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-20T06:42:25Z","lastTransitionTime":"2026-02-20T06:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.829528 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=84.829514209 podStartE2EDuration="1m24.829514209s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:23.75010836 +0000 UTC m=+100.521397348" watchObservedRunningTime="2026-02-20 06:42:25.829514209 +0000 UTC m=+102.600803187" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.831819 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n"] Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.833580 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.835136 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.835748 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.835914 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.836291 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.934096 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/95cce06f-50d2-459e-9cb6-38e5274f383e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.934127 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/95cce06f-50d2-459e-9cb6-38e5274f383e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.934166 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/95cce06f-50d2-459e-9cb6-38e5274f383e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.934405 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95cce06f-50d2-459e-9cb6-38e5274f383e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:25 crc kubenswrapper[4492]: I0220 06:42:25.934445 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95cce06f-50d2-459e-9cb6-38e5274f383e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.034730 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/95cce06f-50d2-459e-9cb6-38e5274f383e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.034774 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/95cce06f-50d2-459e-9cb6-38e5274f383e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.034794 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/95cce06f-50d2-459e-9cb6-38e5274f383e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.034800 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/95cce06f-50d2-459e-9cb6-38e5274f383e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.034829 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95cce06f-50d2-459e-9cb6-38e5274f383e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.034836 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/95cce06f-50d2-459e-9cb6-38e5274f383e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.034843 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95cce06f-50d2-459e-9cb6-38e5274f383e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.035665 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/95cce06f-50d2-459e-9cb6-38e5274f383e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.039324 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95cce06f-50d2-459e-9cb6-38e5274f383e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.047143 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/95cce06f-50d2-459e-9cb6-38e5274f383e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-89f2n\" (UID: \"95cce06f-50d2-459e-9cb6-38e5274f383e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.144310 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.555990 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.556006 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.555990 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:26 crc kubenswrapper[4492]: E0220 06:42:26.556087 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:26 crc kubenswrapper[4492]: E0220 06:42:26.556173 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:26 crc kubenswrapper[4492]: E0220 06:42:26.556228 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.576281 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 03:09:49.34194108 +0000 UTC Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.576326 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.581833 4492 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.973854 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" event={"ID":"95cce06f-50d2-459e-9cb6-38e5274f383e","Type":"ContainerStarted","Data":"78197654d5b396b7efd9318c841d0505ea8961eeb9a1885971452ba207f3c6f0"} Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.974089 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" event={"ID":"95cce06f-50d2-459e-9cb6-38e5274f383e","Type":"ContainerStarted","Data":"1d40d3cede6f9f5f029882883fb7e1c4e4b81c8175ed33696a9e6fbbbbdbd94b"} Feb 20 06:42:26 crc kubenswrapper[4492]: I0220 06:42:26.985173 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-89f2n" podStartSLOduration=85.985161969 podStartE2EDuration="1m25.985161969s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:26.98480722 +0000 UTC m=+103.756096197" watchObservedRunningTime="2026-02-20 06:42:26.985161969 +0000 UTC m=+103.756450947" Feb 20 06:42:27 crc kubenswrapper[4492]: I0220 06:42:27.556675 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:27 crc kubenswrapper[4492]: E0220 06:42:27.556850 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:28 crc kubenswrapper[4492]: I0220 06:42:28.556799 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:28 crc kubenswrapper[4492]: I0220 06:42:28.556796 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:28 crc kubenswrapper[4492]: E0220 06:42:28.557525 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:28 crc kubenswrapper[4492]: I0220 06:42:28.556879 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:28 crc kubenswrapper[4492]: E0220 06:42:28.557578 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:28 crc kubenswrapper[4492]: E0220 06:42:28.557831 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:29 crc kubenswrapper[4492]: I0220 06:42:29.556133 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:29 crc kubenswrapper[4492]: E0220 06:42:29.556322 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:30 crc kubenswrapper[4492]: I0220 06:42:30.556101 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:30 crc kubenswrapper[4492]: I0220 06:42:30.556106 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:30 crc kubenswrapper[4492]: E0220 06:42:30.556338 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:30 crc kubenswrapper[4492]: E0220 06:42:30.556227 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:30 crc kubenswrapper[4492]: I0220 06:42:30.556621 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:30 crc kubenswrapper[4492]: E0220 06:42:30.556769 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:31 crc kubenswrapper[4492]: I0220 06:42:31.556884 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:31 crc kubenswrapper[4492]: E0220 06:42:31.557265 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:31 crc kubenswrapper[4492]: I0220 06:42:31.557408 4492 scope.go:117] "RemoveContainer" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" Feb 20 06:42:31 crc kubenswrapper[4492]: E0220 06:42:31.557582 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f68mj_openshift-ovn-kubernetes(28d6c67f-c4cd-4692-b490-b2b884c72db6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" Feb 20 06:42:32 crc kubenswrapper[4492]: I0220 06:42:32.556275 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:32 crc kubenswrapper[4492]: I0220 06:42:32.556299 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:32 crc kubenswrapper[4492]: I0220 06:42:32.556329 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:32 crc kubenswrapper[4492]: E0220 06:42:32.556360 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:32 crc kubenswrapper[4492]: E0220 06:42:32.556426 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:32 crc kubenswrapper[4492]: E0220 06:42:32.556549 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:33 crc kubenswrapper[4492]: I0220 06:42:33.556164 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:33 crc kubenswrapper[4492]: E0220 06:42:33.557313 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.556856 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.557350 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.557382 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:34 crc kubenswrapper[4492]: E0220 06:42:34.557440 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:34 crc kubenswrapper[4492]: E0220 06:42:34.557756 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:34 crc kubenswrapper[4492]: E0220 06:42:34.557861 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.992032 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/1.log" Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.992489 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/0.log" Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.992533 4492 generic.go:334] "Generic (PLEG): container finished" podID="dcc11973-022a-47f9-b1e2-23e945352e7f" containerID="d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8" exitCode=1 Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.992562 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerDied","Data":"d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8"} Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.992594 4492 scope.go:117] "RemoveContainer" containerID="d1f6d8bf1d85811b2c7d87c456b0fa6082e35ac1bb207821aec2bf6230d576d8" Feb 20 06:42:34 crc kubenswrapper[4492]: I0220 06:42:34.993073 4492 scope.go:117] "RemoveContainer" containerID="d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8" Feb 20 06:42:34 crc kubenswrapper[4492]: E0220 06:42:34.993295 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2rr9j_openshift-multus(dcc11973-022a-47f9-b1e2-23e945352e7f)\"" pod="openshift-multus/multus-2rr9j" podUID="dcc11973-022a-47f9-b1e2-23e945352e7f" Feb 20 06:42:35 crc kubenswrapper[4492]: I0220 06:42:35.556087 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:35 crc kubenswrapper[4492]: E0220 06:42:35.556193 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:35 crc kubenswrapper[4492]: I0220 06:42:35.997440 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/1.log" Feb 20 06:42:36 crc kubenswrapper[4492]: I0220 06:42:36.555916 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:36 crc kubenswrapper[4492]: I0220 06:42:36.555933 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:36 crc kubenswrapper[4492]: I0220 06:42:36.555973 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:36 crc kubenswrapper[4492]: E0220 06:42:36.556058 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:36 crc kubenswrapper[4492]: E0220 06:42:36.556130 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:36 crc kubenswrapper[4492]: E0220 06:42:36.556200 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:37 crc kubenswrapper[4492]: I0220 06:42:37.556690 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:37 crc kubenswrapper[4492]: E0220 06:42:37.556811 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:38 crc kubenswrapper[4492]: I0220 06:42:38.556119 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:38 crc kubenswrapper[4492]: I0220 06:42:38.556179 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:38 crc kubenswrapper[4492]: E0220 06:42:38.556226 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:38 crc kubenswrapper[4492]: E0220 06:42:38.556265 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:38 crc kubenswrapper[4492]: I0220 06:42:38.556134 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:38 crc kubenswrapper[4492]: E0220 06:42:38.556600 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:39 crc kubenswrapper[4492]: I0220 06:42:39.556222 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:39 crc kubenswrapper[4492]: E0220 06:42:39.556353 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:40 crc kubenswrapper[4492]: I0220 06:42:40.556845 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:40 crc kubenswrapper[4492]: I0220 06:42:40.556885 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:40 crc kubenswrapper[4492]: E0220 06:42:40.556929 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:40 crc kubenswrapper[4492]: E0220 06:42:40.556982 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:40 crc kubenswrapper[4492]: I0220 06:42:40.557426 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:40 crc kubenswrapper[4492]: E0220 06:42:40.557867 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:41 crc kubenswrapper[4492]: I0220 06:42:41.556453 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:41 crc kubenswrapper[4492]: E0220 06:42:41.556625 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:42 crc kubenswrapper[4492]: I0220 06:42:42.555878 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:42 crc kubenswrapper[4492]: E0220 06:42:42.556424 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:42 crc kubenswrapper[4492]: I0220 06:42:42.555990 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:42 crc kubenswrapper[4492]: E0220 06:42:42.556660 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:42 crc kubenswrapper[4492]: I0220 06:42:42.555948 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:42 crc kubenswrapper[4492]: E0220 06:42:42.556862 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:43 crc kubenswrapper[4492]: E0220 06:42:43.542417 4492 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 20 06:42:43 crc kubenswrapper[4492]: I0220 06:42:43.557984 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:43 crc kubenswrapper[4492]: E0220 06:42:43.558089 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:43 crc kubenswrapper[4492]: E0220 06:42:43.623241 4492 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 06:42:44 crc kubenswrapper[4492]: I0220 06:42:44.556671 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:44 crc kubenswrapper[4492]: I0220 06:42:44.556695 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:44 crc kubenswrapper[4492]: I0220 06:42:44.556728 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:44 crc kubenswrapper[4492]: E0220 06:42:44.556782 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:44 crc kubenswrapper[4492]: E0220 06:42:44.556842 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:44 crc kubenswrapper[4492]: E0220 06:42:44.556903 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:45 crc kubenswrapper[4492]: I0220 06:42:45.556534 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:45 crc kubenswrapper[4492]: E0220 06:42:45.556635 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:46 crc kubenswrapper[4492]: I0220 06:42:46.556317 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:46 crc kubenswrapper[4492]: I0220 06:42:46.556317 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:46 crc kubenswrapper[4492]: I0220 06:42:46.556365 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:46 crc kubenswrapper[4492]: I0220 06:42:46.556791 4492 scope.go:117] "RemoveContainer" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" Feb 20 06:42:46 crc kubenswrapper[4492]: E0220 06:42:46.556956 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:46 crc kubenswrapper[4492]: E0220 06:42:46.557048 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:46 crc kubenswrapper[4492]: E0220 06:42:46.557226 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:47 crc kubenswrapper[4492]: I0220 06:42:47.026681 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/3.log" Feb 20 06:42:47 crc kubenswrapper[4492]: I0220 06:42:47.029204 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerStarted","Data":"74a35c1868a572cbfcd7507555baf61f36d97d28af5ac21d481547bd4c44f779"} Feb 20 06:42:47 crc kubenswrapper[4492]: I0220 06:42:47.029606 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:42:47 crc kubenswrapper[4492]: I0220 06:42:47.052235 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podStartSLOduration=106.052200689 podStartE2EDuration="1m46.052200689s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:42:47.051857521 +0000 UTC m=+123.823146500" watchObservedRunningTime="2026-02-20 06:42:47.052200689 +0000 UTC m=+123.823489667" Feb 20 06:42:47 crc kubenswrapper[4492]: I0220 06:42:47.336879 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p9vbm"] Feb 20 06:42:47 crc kubenswrapper[4492]: I0220 06:42:47.336970 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:47 crc kubenswrapper[4492]: E0220 06:42:47.337055 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:47 crc kubenswrapper[4492]: I0220 06:42:47.556154 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:47 crc kubenswrapper[4492]: E0220 06:42:47.556281 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:48 crc kubenswrapper[4492]: I0220 06:42:48.556265 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:48 crc kubenswrapper[4492]: E0220 06:42:48.556407 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:48 crc kubenswrapper[4492]: I0220 06:42:48.556623 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:48 crc kubenswrapper[4492]: E0220 06:42:48.556671 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:48 crc kubenswrapper[4492]: E0220 06:42:48.624062 4492 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 20 06:42:49 crc kubenswrapper[4492]: I0220 06:42:49.556864 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:49 crc kubenswrapper[4492]: I0220 06:42:49.556918 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:49 crc kubenswrapper[4492]: E0220 06:42:49.557058 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:49 crc kubenswrapper[4492]: I0220 06:42:49.557234 4492 scope.go:117] "RemoveContainer" containerID="d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8" Feb 20 06:42:49 crc kubenswrapper[4492]: E0220 06:42:49.557355 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:50 crc kubenswrapper[4492]: I0220 06:42:50.039448 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/1.log" Feb 20 06:42:50 crc kubenswrapper[4492]: I0220 06:42:50.039532 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerStarted","Data":"e63ed096fe97531f09118ad4101afeac3694e124f9060047cedcde53421e659a"} Feb 20 06:42:50 crc kubenswrapper[4492]: I0220 06:42:50.556539 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:50 crc kubenswrapper[4492]: E0220 06:42:50.556648 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:50 crc kubenswrapper[4492]: I0220 06:42:50.556539 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:50 crc kubenswrapper[4492]: E0220 06:42:50.556815 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:51 crc kubenswrapper[4492]: I0220 06:42:51.556783 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:51 crc kubenswrapper[4492]: I0220 06:42:51.556856 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:51 crc kubenswrapper[4492]: E0220 06:42:51.556933 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:51 crc kubenswrapper[4492]: E0220 06:42:51.557062 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:52 crc kubenswrapper[4492]: I0220 06:42:52.138270 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:42:52 crc kubenswrapper[4492]: I0220 06:42:52.555852 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:52 crc kubenswrapper[4492]: I0220 06:42:52.555894 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:52 crc kubenswrapper[4492]: E0220 06:42:52.555978 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 20 06:42:52 crc kubenswrapper[4492]: E0220 06:42:52.556043 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 20 06:42:53 crc kubenswrapper[4492]: I0220 06:42:53.556722 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:53 crc kubenswrapper[4492]: E0220 06:42:53.557124 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 20 06:42:53 crc kubenswrapper[4492]: I0220 06:42:53.557531 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:53 crc kubenswrapper[4492]: E0220 06:42:53.557635 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p9vbm" podUID="0825f068-2c78-44a0-8f24-679e4521f069" Feb 20 06:42:54 crc kubenswrapper[4492]: I0220 06:42:54.556467 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:42:54 crc kubenswrapper[4492]: I0220 06:42:54.556501 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:42:54 crc kubenswrapper[4492]: I0220 06:42:54.558303 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 06:42:54 crc kubenswrapper[4492]: I0220 06:42:54.561260 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 06:42:55 crc kubenswrapper[4492]: I0220 06:42:55.556660 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:42:55 crc kubenswrapper[4492]: I0220 06:42:55.556791 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:42:55 crc kubenswrapper[4492]: I0220 06:42:55.558777 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 06:42:55 crc kubenswrapper[4492]: I0220 06:42:55.560141 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 06:42:55 crc kubenswrapper[4492]: I0220 06:42:55.560652 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 06:42:55 crc kubenswrapper[4492]: I0220 06:42:55.560777 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.769189 4492 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.802490 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t8jr5"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.803116 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.803830 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.804347 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.812687 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.814781 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.814883 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.815034 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.815203 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.815283 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.815244 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.815256 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.816073 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.816614 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.816614 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.816771 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.816689 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.817849 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7ggr"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.818251 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.827828 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.827835 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.828522 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.828612 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.828628 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.828833 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.829383 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.829850 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.829958 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.830406 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.830649 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.830729 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.830748 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.831337 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.831486 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.831641 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.832204 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.833309 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.833560 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840009 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840022 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840101 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840123 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840244 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840787 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840813 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.840930 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841009 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841168 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841329 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841395 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841498 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841528 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841724 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841866 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841932 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.841979 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.842166 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.842269 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.842287 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.842270 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.842185 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.842412 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.842837 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gtnl8"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.843273 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.843600 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-n62tv"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.843848 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-n62tv" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.844246 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hr8jm"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.844560 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.844892 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p8dlg"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.845420 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.845733 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sjzvg"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.846295 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.846619 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.846815 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.846838 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wbkxj"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.847250 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.847693 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.848120 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.851298 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.854773 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.864402 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.868817 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.876668 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.876897 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877067 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877172 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877238 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877075 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877376 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877463 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877533 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877643 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877801 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877935 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.877954 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.878069 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.878447 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.878523 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.878596 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.878808 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.879271 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b8qcp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.879576 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.879838 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.880017 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.879280 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.879356 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.880346 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.879462 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.880664 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.880806 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.880838 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.880965 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.880982 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881039 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2w8gp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881136 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881409 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881712 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881781 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881832 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881849 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881905 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.881920 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.882247 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.882335 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gblsk"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.882911 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.882963 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.883380 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.883525 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.883807 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884156 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vt7z5"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884226 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884302 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884520 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884584 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884624 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884526 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884592 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.884768 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.885577 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.886450 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889838 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nth9k\" (UniqueName: \"kubernetes.io/projected/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-kube-api-access-nth9k\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889869 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889879 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c3dd185-faf8-4150-81b6-ef352c48148b-images\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889902 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-trusted-ca-bundle\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889921 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-serving-cert\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889938 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/471752d9-2f13-410f-831c-4785c2813fba-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889954 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-config\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889970 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.889990 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-oauth-serving-cert\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890024 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-trusted-ca\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890064 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kth7x\" (UniqueName: \"kubernetes.io/projected/327efe9d-4122-45d6-ad6b-62e41ae73a74-kube-api-access-kth7x\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890087 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/123421da-365d-42ef-a5ef-1a129ec42490-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890103 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890125 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-encryption-config\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890140 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890162 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de123619-8818-4fea-a9d0-75e498b66cca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-66mcq\" (UID: \"de123619-8818-4fea-a9d0-75e498b66cca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890178 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c3dd185-faf8-4150-81b6-ef352c48148b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890192 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-oauth-config\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890208 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890223 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890254 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890274 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/327efe9d-4122-45d6-ad6b-62e41ae73a74-serving-cert\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890289 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-machine-approver-tls\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890302 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-audit-policies\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890318 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnkt5\" (UniqueName: \"kubernetes.io/projected/de123619-8818-4fea-a9d0-75e498b66cca-kube-api-access-fnkt5\") pod \"cluster-samples-operator-665b6dd947-66mcq\" (UID: \"de123619-8818-4fea-a9d0-75e498b66cca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890346 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l6kp\" (UniqueName: \"kubernetes.io/projected/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-kube-api-access-6l6kp\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890380 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxngm\" (UniqueName: \"kubernetes.io/projected/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-kube-api-access-pxngm\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890737 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-client-ca\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890774 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890803 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-etcd-serving-ca\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890841 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-config\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890873 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6twn4\" (UniqueName: \"kubernetes.io/projected/80f87acb-f24f-4f79-9677-227c929e0497-kube-api-access-6twn4\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.890901 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891000 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123421da-365d-42ef-a5ef-1a129ec42490-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891032 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx8r2\" (UniqueName: \"kubernetes.io/projected/8f9c533a-3aa4-4413-914d-6d24f234c092-kube-api-access-tx8r2\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891059 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-dir\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891110 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891136 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891162 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/471752d9-2f13-410f-831c-4785c2813fba-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891186 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxvd8\" (UniqueName: \"kubernetes.io/projected/471752d9-2f13-410f-831c-4785c2813fba-kube-api-access-gxvd8\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891252 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5559ffd4-7a7c-4603-b80e-0ea3159da609-node-pullsecrets\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891277 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-image-import-ca\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891311 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c3dd185-faf8-4150-81b6-ef352c48148b-config\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891348 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-serving-cert\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891384 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-serving-cert\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891417 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-config\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891434 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-etcd-client\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891459 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-policies\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891499 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891517 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891538 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f9c533a-3aa4-4413-914d-6d24f234c092-serving-cert\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891553 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891571 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-serving-cert\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891590 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-service-ca-bundle\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891604 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mskv\" (UniqueName: \"kubernetes.io/projected/123421da-365d-42ef-a5ef-1a129ec42490-kube-api-access-8mskv\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891669 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-service-ca\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891688 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m7k6\" (UniqueName: \"kubernetes.io/projected/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-kube-api-access-6m7k6\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891704 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891721 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891735 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-audit-dir\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891775 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891796 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891809 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-config\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891830 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/471752d9-2f13-410f-831c-4785c2813fba-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891862 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891878 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891892 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-config\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891911 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-proxy-tls\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891925 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-serving-cert\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891962 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80f87acb-f24f-4f79-9677-227c929e0497-serving-cert\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891980 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5559ffd4-7a7c-4603-b80e-0ea3159da609-audit-dir\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.891993 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-config\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892012 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-etcd-client\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892026 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-encryption-config\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892067 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czmrj\" (UniqueName: \"kubernetes.io/projected/4c3dd185-faf8-4150-81b6-ef352c48148b-kube-api-access-czmrj\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892087 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-config\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892118 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-audit\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892136 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-auth-proxy-config\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892154 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ff9t\" (UniqueName: \"kubernetes.io/projected/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-kube-api-access-8ff9t\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892168 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnlsx\" (UniqueName: \"kubernetes.io/projected/5559ffd4-7a7c-4603-b80e-0ea3159da609-kube-api-access-vnlsx\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892192 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892209 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87fn\" (UniqueName: \"kubernetes.io/projected/ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4-kube-api-access-q87fn\") pod \"downloads-7954f5f757-n62tv\" (UID: \"ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4\") " pod="openshift-console/downloads-7954f5f757-n62tv" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892263 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-client-ca\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892279 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8rhf\" (UniqueName: \"kubernetes.io/projected/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-kube-api-access-b8rhf\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892293 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-images\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892309 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9sth\" (UniqueName: \"kubernetes.io/projected/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-kube-api-access-s9sth\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.892365 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p48zj\" (UniqueName: \"kubernetes.io/projected/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-kube-api-access-p48zj\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.896690 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.898191 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.901254 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t9jrw"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.919062 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.920900 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.920937 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.922628 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.923102 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.925773 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5snjq"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.927246 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.925784 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.926955 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.927092 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.932620 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.933695 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.934066 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.935798 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.935952 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.938921 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.939118 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.939273 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.939627 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.939747 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.939829 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.939888 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.939668 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-q28nt"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.941135 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.941822 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.941974 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.942199 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.942649 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.942833 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.943380 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.943833 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.945089 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.946012 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.946184 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.946429 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vks6z"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.946685 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.946826 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.947616 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7ggr"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.948501 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.952418 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.952590 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.958399 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.959261 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.960455 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p8dlg"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.962153 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-n62tv"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.963424 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.963860 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wbkxj"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.965699 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hr8jm"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.965795 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.966436 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.967353 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bmc25"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.968256 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.968544 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sjzvg"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.969315 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.970190 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2w8gp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.971027 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vt7z5"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.972498 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.972967 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.975444 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gtnl8"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.976696 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t9jrw"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.977131 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.977931 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b8qcp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.978667 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gblsk"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.979384 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.980318 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.983451 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.991600 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993284 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p48zj\" (UniqueName: \"kubernetes.io/projected/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-kube-api-access-p48zj\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993328 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-serving-cert\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993352 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nth9k\" (UniqueName: \"kubernetes.io/projected/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-kube-api-access-nth9k\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993374 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c3dd185-faf8-4150-81b6-ef352c48148b-images\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993415 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-trusted-ca-bundle\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993449 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-trusted-ca\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993467 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/471752d9-2f13-410f-831c-4785c2813fba-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993531 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-config\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993553 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993638 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-oauth-serving-cert\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993664 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kth7x\" (UniqueName: \"kubernetes.io/projected/327efe9d-4122-45d6-ad6b-62e41ae73a74-kube-api-access-kth7x\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993724 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/123421da-365d-42ef-a5ef-1a129ec42490-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993746 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de123619-8818-4fea-a9d0-75e498b66cca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-66mcq\" (UID: \"de123619-8818-4fea-a9d0-75e498b66cca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993803 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.993852 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7923524-5488-4b05-a52f-7b198e0ba4ad-metrics-tls\") pod \"dns-operator-744455d44c-vt7z5\" (UID: \"c7923524-5488-4b05-a52f-7b198e0ba4ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994142 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-encryption-config\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994175 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994544 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c3dd185-faf8-4150-81b6-ef352c48148b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994586 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-oauth-config\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994607 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/327efe9d-4122-45d6-ad6b-62e41ae73a74-serving-cert\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994630 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994652 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994673 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994693 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-machine-approver-tls\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994711 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-audit-policies\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994731 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnkt5\" (UniqueName: \"kubernetes.io/projected/de123619-8818-4fea-a9d0-75e498b66cca-kube-api-access-fnkt5\") pod \"cluster-samples-operator-665b6dd947-66mcq\" (UID: \"de123619-8818-4fea-a9d0-75e498b66cca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994751 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l6kp\" (UniqueName: \"kubernetes.io/projected/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-kube-api-access-6l6kp\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994771 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxngm\" (UniqueName: \"kubernetes.io/projected/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-kube-api-access-pxngm\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994792 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-client-ca\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994814 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994833 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994874 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-etcd-serving-ca\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994896 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-config\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994939 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6twn4\" (UniqueName: \"kubernetes.io/projected/80f87acb-f24f-4f79-9677-227c929e0497-kube-api-access-6twn4\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994960 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.994981 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123421da-365d-42ef-a5ef-1a129ec42490-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995001 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx8r2\" (UniqueName: \"kubernetes.io/projected/8f9c533a-3aa4-4413-914d-6d24f234c092-kube-api-access-tx8r2\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995017 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-dir\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995038 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995059 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/471752d9-2f13-410f-831c-4785c2813fba-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995078 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxvd8\" (UniqueName: \"kubernetes.io/projected/471752d9-2f13-410f-831c-4785c2813fba-kube-api-access-gxvd8\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995096 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5559ffd4-7a7c-4603-b80e-0ea3159da609-node-pullsecrets\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995146 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-serving-cert\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995172 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-trusted-ca\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995197 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-image-import-ca\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995222 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c3dd185-faf8-4150-81b6-ef352c48148b-config\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995242 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-serving-cert\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995259 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-etcd-client\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995283 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hdsk\" (UniqueName: \"kubernetes.io/projected/c7923524-5488-4b05-a52f-7b198e0ba4ad-kube-api-access-8hdsk\") pod \"dns-operator-744455d44c-vt7z5\" (UID: \"c7923524-5488-4b05-a52f-7b198e0ba4ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995305 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-config\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995334 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995352 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-policies\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995371 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995390 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-serving-cert\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995410 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f9c533a-3aa4-4413-914d-6d24f234c092-serving-cert\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995428 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995445 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-service-ca-bundle\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995464 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mskv\" (UniqueName: \"kubernetes.io/projected/123421da-365d-42ef-a5ef-1a129ec42490-kube-api-access-8mskv\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995509 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-service-ca\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995531 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m7k6\" (UniqueName: \"kubernetes.io/projected/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-kube-api-access-6m7k6\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995553 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995572 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995593 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995609 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-audit-dir\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995636 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995655 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995677 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-config\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995698 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/471752d9-2f13-410f-831c-4785c2813fba-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995716 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995735 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80f87acb-f24f-4f79-9677-227c929e0497-serving-cert\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995750 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-config\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995767 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-proxy-tls\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995784 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-serving-cert\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995800 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5559ffd4-7a7c-4603-b80e-0ea3159da609-audit-dir\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995814 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-config\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995832 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-etcd-client\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995931 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5snjq"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.995960 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.996622 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.996673 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl"] Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.997369 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-encryption-config\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.997416 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czmrj\" (UniqueName: \"kubernetes.io/projected/4c3dd185-faf8-4150-81b6-ef352c48148b-kube-api-access-czmrj\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.997446 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-config\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.997509 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-audit\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.997508 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c3dd185-faf8-4150-81b6-ef352c48148b-images\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.997932 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-etcd-serving-ca\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.998071 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-config\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.998178 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-auth-proxy-config\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.998202 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-config\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.998231 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ff9t\" (UniqueName: \"kubernetes.io/projected/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-kube-api-access-8ff9t\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.998262 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnlsx\" (UniqueName: \"kubernetes.io/projected/5559ffd4-7a7c-4603-b80e-0ea3159da609-kube-api-access-vnlsx\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.998288 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.999054 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-trusted-ca-bundle\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.998338 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87fn\" (UniqueName: \"kubernetes.io/projected/ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4-kube-api-access-q87fn\") pod \"downloads-7954f5f757-n62tv\" (UID: \"ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4\") " pod="openshift-console/downloads-7954f5f757-n62tv" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.999398 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-images\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.999431 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-client-ca\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.999449 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8rhf\" (UniqueName: \"kubernetes.io/projected/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-kube-api-access-b8rhf\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.999487 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9sth\" (UniqueName: \"kubernetes.io/projected/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-kube-api-access-s9sth\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.999854 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/123421da-365d-42ef-a5ef-1a129ec42490-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:56 crc kubenswrapper[4492]: I0220 06:42:56.999873 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-config\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.001909 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bmc25"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.001990 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-auth-proxy-config\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.002575 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-client-ca\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.002674 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-images\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.003455 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-oauth-serving-cert\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.003603 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-audit\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.003849 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-encryption-config\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.004274 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-serving-cert\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.005368 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-56v6b"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.006221 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.006243 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.005450 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-dir\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.005418 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-service-ca-bundle\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.006349 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.007062 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.007660 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-service-ca\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.008041 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.009740 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.009873 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-policies\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.010210 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-config\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.012579 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-etcd-client\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.014527 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/471752d9-2f13-410f-831c-4785c2813fba-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.014625 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5559ffd4-7a7c-4603-b80e-0ea3159da609-node-pullsecrets\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.016110 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.016465 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-serving-cert\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.016563 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.017038 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/123421da-365d-42ef-a5ef-1a129ec42490-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.017427 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-etcd-client\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.017596 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-image-import-ca\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.018085 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c3dd185-faf8-4150-81b6-ef352c48148b-config\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.018610 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.018907 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.019167 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.020292 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.021525 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.021689 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.021748 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-audit-dir\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.024026 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.024290 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.024346 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.024348 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.024758 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.027035 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-serving-cert\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.027238 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.027353 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-56v6b"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.027389 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vks6z"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.027402 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t8jr5"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.027803 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/327efe9d-4122-45d6-ad6b-62e41ae73a74-config\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.028528 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f9c533a-3aa4-4413-914d-6d24f234c092-serving-cert\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.028647 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-encryption-config\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.028729 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/471752d9-2f13-410f-831c-4785c2813fba-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.028765 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/de123619-8818-4fea-a9d0-75e498b66cca-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-66mcq\" (UID: \"de123619-8818-4fea-a9d0-75e498b66cca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.029203 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.029519 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-config\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.029544 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.030522 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.030549 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-psfnh"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.031114 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7sf2k"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.031942 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/327efe9d-4122-45d6-ad6b-62e41ae73a74-serving-cert\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.032076 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.032214 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-machine-approver-tls\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.032991 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-audit-policies\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.033188 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.033297 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7sf2k"] Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.033362 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.033646 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5559ffd4-7a7c-4603-b80e-0ea3159da609-audit-dir\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.033744 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-client-ca\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.033896 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80f87acb-f24f-4f79-9677-227c929e0497-serving-cert\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.034381 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5559ffd4-7a7c-4603-b80e-0ea3159da609-trusted-ca-bundle\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.034565 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-config\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.035011 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-serving-cert\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.036695 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.037622 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5559ffd4-7a7c-4603-b80e-0ea3159da609-serving-cert\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.038737 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-oauth-config\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.038872 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.039875 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c3dd185-faf8-4150-81b6-ef352c48148b-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.040512 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-proxy-tls\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.045621 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.066391 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.085181 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.100093 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hdsk\" (UniqueName: \"kubernetes.io/projected/c7923524-5488-4b05-a52f-7b198e0ba4ad-kube-api-access-8hdsk\") pod \"dns-operator-744455d44c-vt7z5\" (UID: \"c7923524-5488-4b05-a52f-7b198e0ba4ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.100245 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7923524-5488-4b05-a52f-7b198e0ba4ad-metrics-tls\") pod \"dns-operator-744455d44c-vt7z5\" (UID: \"c7923524-5488-4b05-a52f-7b198e0ba4ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.105658 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.124821 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.150515 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.164945 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.184879 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.225236 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.245777 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.266094 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.286365 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.305509 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.325170 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.345125 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.366524 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.385719 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.406023 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.425883 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.455109 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.465580 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.485957 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.505746 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.525433 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.545584 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.564672 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.585231 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.604931 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.625921 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.634315 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c7923524-5488-4b05-a52f-7b198e0ba4ad-metrics-tls\") pod \"dns-operator-744455d44c-vt7z5\" (UID: \"c7923524-5488-4b05-a52f-7b198e0ba4ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.645380 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.665889 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.684876 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.705256 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.726131 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.745377 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.765516 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.785654 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.805033 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.825409 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.844967 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.874597 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.885268 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.905533 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.924850 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.944108 4492 request.go:700] Waited for 1.011347357s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.945509 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.964882 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 06:42:57 crc kubenswrapper[4492]: I0220 06:42:57.985637 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.005127 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.024908 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.045847 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.065097 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.085279 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.104967 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.125192 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.144784 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.165087 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.185141 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.205655 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.224994 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.244946 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.265587 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.285445 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.305051 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.324708 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.344611 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.365281 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.384900 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.404631 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.425036 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.445827 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.465666 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.485469 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.504806 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.524927 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.544863 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.565588 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.605643 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.625206 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.645361 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.677415 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p48zj\" (UniqueName: \"kubernetes.io/projected/4ce7edf9-9ce6-4c47-ac39-3881c7f1a814-kube-api-access-p48zj\") pod \"openshift-apiserver-operator-796bbdcf4f-5clmz\" (UID: \"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.697006 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nth9k\" (UniqueName: \"kubernetes.io/projected/6fc4232f-8bca-43c5-a01e-02d2b58f43c9-kube-api-access-nth9k\") pod \"apiserver-7bbb656c7d-t7xh6\" (UID: \"6fc4232f-8bca-43c5-a01e-02d2b58f43c9\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.717499 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6twn4\" (UniqueName: \"kubernetes.io/projected/80f87acb-f24f-4f79-9677-227c929e0497-kube-api-access-6twn4\") pod \"route-controller-manager-6576b87f9c-v2kxs\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.727678 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.741133 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/471752d9-2f13-410f-831c-4785c2813fba-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.757385 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czmrj\" (UniqueName: \"kubernetes.io/projected/4c3dd185-faf8-4150-81b6-ef352c48148b-kube-api-access-czmrj\") pod \"machine-api-operator-5694c8668f-sjzvg\" (UID: \"4c3dd185-faf8-4150-81b6-ef352c48148b\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.776206 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87fn\" (UniqueName: \"kubernetes.io/projected/ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4-kube-api-access-q87fn\") pod \"downloads-7954f5f757-n62tv\" (UID: \"ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4\") " pod="openshift-console/downloads-7954f5f757-n62tv" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.796815 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ff9t\" (UniqueName: \"kubernetes.io/projected/3be5a61d-eaaa-4eb7-bdc3-50d67493a117-kube-api-access-8ff9t\") pod \"machine-approver-56656f9798-mfr74\" (UID: \"3be5a61d-eaaa-4eb7-bdc3-50d67493a117\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.798591 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.818979 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnlsx\" (UniqueName: \"kubernetes.io/projected/5559ffd4-7a7c-4603-b80e-0ea3159da609-kube-api-access-vnlsx\") pod \"apiserver-76f77b778f-t8jr5\" (UID: \"5559ffd4-7a7c-4603-b80e-0ea3159da609\") " pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.836654 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9sth\" (UniqueName: \"kubernetes.io/projected/b5cd6268-e4fa-47ae-9cac-d03da9e84fe7-kube-api-access-s9sth\") pod \"openshift-config-operator-7777fb866f-xfgjc\" (UID: \"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.857323 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx8r2\" (UniqueName: \"kubernetes.io/projected/8f9c533a-3aa4-4413-914d-6d24f234c092-kube-api-access-tx8r2\") pod \"controller-manager-879f6c89f-v7ggr\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.879325 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8rhf\" (UniqueName: \"kubernetes.io/projected/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-kube-api-access-b8rhf\") pod \"console-f9d7485db-wbkxj\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.890525 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz"] Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.903980 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kth7x\" (UniqueName: \"kubernetes.io/projected/327efe9d-4122-45d6-ad6b-62e41ae73a74-kube-api-access-kth7x\") pod \"authentication-operator-69f744f599-hr8jm\" (UID: \"327efe9d-4122-45d6-ad6b-62e41ae73a74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.905543 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.916761 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.926240 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.927060 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.943091 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.944850 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.955952 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.965108 4492 request.go:700] Waited for 1.958435179s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.970072 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sjzvg"] Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.970292 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 06:42:58 crc kubenswrapper[4492]: W0220 06:42:58.978657 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c3dd185_faf8_4150_81b6_ef352c48148b.slice/crio-c6bf4efec9691cca7640d45fc7bc16b6821940247bd22719bfa7d279ed8bf578 WatchSource:0}: Error finding container c6bf4efec9691cca7640d45fc7bc16b6821940247bd22719bfa7d279ed8bf578: Status 404 returned error can't find the container with id c6bf4efec9691cca7640d45fc7bc16b6821940247bd22719bfa7d279ed8bf578 Feb 20 06:42:58 crc kubenswrapper[4492]: I0220 06:42:58.999237 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.008706 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mskv\" (UniqueName: \"kubernetes.io/projected/123421da-365d-42ef-a5ef-1a129ec42490-kube-api-access-8mskv\") pod \"kube-storage-version-migrator-operator-b67b599dd-jldvk\" (UID: \"123421da-365d-42ef-a5ef-1a129ec42490\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.021115 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m7k6\" (UniqueName: \"kubernetes.io/projected/010dc2a5-404a-47ed-881e-5cb6b5f6bae6-kube-api-access-6m7k6\") pod \"machine-config-operator-74547568cd-qh4bk\" (UID: \"010dc2a5-404a-47ed-881e-5cb6b5f6bae6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.036780 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.046140 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.046318 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxvd8\" (UniqueName: \"kubernetes.io/projected/471752d9-2f13-410f-831c-4785c2813fba-kube-api-access-gxvd8\") pod \"cluster-image-registry-operator-dc59b4c8b-lrqc8\" (UID: \"471752d9-2f13-410f-831c-4785c2813fba\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.066075 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-n62tv" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.070446 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.071849 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" event={"ID":"4c3dd185-faf8-4150-81b6-ef352c48148b","Type":"ContainerStarted","Data":"c6bf4efec9691cca7640d45fc7bc16b6821940247bd22719bfa7d279ed8bf578"} Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.074548 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" event={"ID":"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814","Type":"ContainerStarted","Data":"2b3b4ce6bb2fa0b2add8d9f1287b2c13ea104a40d61a3b60a0bd902701f1fc1a"} Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.074569 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" event={"ID":"4ce7edf9-9ce6-4c47-ac39-3881c7f1a814","Type":"ContainerStarted","Data":"1eafa66fe79e415944af9a3e44a8d3947e21c2c4cce1bdbfea70d320833500ab"} Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.077106 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnkt5\" (UniqueName: \"kubernetes.io/projected/de123619-8818-4fea-a9d0-75e498b66cca-kube-api-access-fnkt5\") pod \"cluster-samples-operator-665b6dd947-66mcq\" (UID: \"de123619-8818-4fea-a9d0-75e498b66cca\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.083095 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.084360 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" event={"ID":"3be5a61d-eaaa-4eb7-bdc3-50d67493a117","Type":"ContainerStarted","Data":"cc3a1b132df136a038f9d208a014a6f6c8f7b690a15812da601f39a97bb08997"} Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.099509 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l6kp\" (UniqueName: \"kubernetes.io/projected/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-kube-api-access-6l6kp\") pod \"oauth-openshift-558db77b4-p8dlg\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.128358 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-t8jr5"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.131082 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.131367 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxngm\" (UniqueName: \"kubernetes.io/projected/6720f917-62d7-49d3-9bfb-ce7f9cfa89a2-kube-api-access-pxngm\") pod \"console-operator-58897d9998-gtnl8\" (UID: \"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2\") " pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.148697 4492 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.166795 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.167019 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.169459 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.170620 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.175743 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.185425 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 06:42:59 crc kubenswrapper[4492]: W0220 06:42:59.199219 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fc4232f_8bca_43c5_a01e_02d2b58f43c9.slice/crio-4c50a4f9cec87e5190a83495525932e40f2d370bc8767ac7f2a2723b1821a681 WatchSource:0}: Error finding container 4c50a4f9cec87e5190a83495525932e40f2d370bc8767ac7f2a2723b1821a681: Status 404 returned error can't find the container with id 4c50a4f9cec87e5190a83495525932e40f2d370bc8767ac7f2a2723b1821a681 Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.206647 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.234745 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7ggr"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.236589 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.255211 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hdsk\" (UniqueName: \"kubernetes.io/projected/c7923524-5488-4b05-a52f-7b198e0ba4ad-kube-api-access-8hdsk\") pod \"dns-operator-744455d44c-vt7z5\" (UID: \"c7923524-5488-4b05-a52f-7b198e0ba4ad\") " pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327708 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f28t2\" (UniqueName: \"kubernetes.io/projected/11f1e06a-5f7f-413c-8682-62a523068334-kube-api-access-f28t2\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327743 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb4a9851-d201-4a71-b26f-fff01b3c0e99-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327773 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5e1e31e-b55a-4db6-8382-ede623b580d7-metrics-tls\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327793 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/58a0e454-50cb-4540-aeb6-b66435322029-webhook-cert\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327813 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2bc1973-e22e-47e6-88bf-edab0829592e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zc5kp\" (UID: \"a2bc1973-e22e-47e6-88bf-edab0829592e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327838 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccz8g\" (UniqueName: \"kubernetes.io/projected/58a0e454-50cb-4540-aeb6-b66435322029-kube-api-access-ccz8g\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327855 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59681aac-a82c-44a9-80bc-0870b8b51905-config\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327873 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e50b25-8add-403e-bf95-b4825c548852-proxy-tls\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327899 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8hcs\" (UniqueName: \"kubernetes.io/projected/c5e1e31e-b55a-4db6-8382-ede623b580d7-kube-api-access-l8hcs\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327916 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjr84\" (UniqueName: \"kubernetes.io/projected/60bf2317-c2a3-4703-8e92-1c8d7c25a857-kube-api-access-mjr84\") pod \"package-server-manager-789f6589d5-jb6vd\" (UID: \"60bf2317-c2a3-4703-8e92-1c8d7c25a857\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327938 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.327986 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp4x4\" (UniqueName: \"kubernetes.io/projected/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-kube-api-access-xp4x4\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.328075 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-ca\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.328106 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-certificates\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.328158 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f1e06a-5f7f-413c-8682-62a523068334-serving-cert\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.328195 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2fcb6c86-c19d-4775-8a95-29e3d65b3926-signing-key\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.328654 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-serving-cert\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.329244 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:42:59.829226797 +0000 UTC m=+136.600515775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.330249 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-config\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.330757 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.330882 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txwb6\" (UniqueName: \"kubernetes.io/projected/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-kube-api-access-txwb6\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331413 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/60bf2317-c2a3-4703-8e92-1c8d7c25a857-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jb6vd\" (UID: \"60bf2317-c2a3-4703-8e92-1c8d7c25a857\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331446 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5e81408f-b0aa-4bde-be40-d132cd172ec9-srv-cert\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331506 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2fcb6c86-c19d-4775-8a95-29e3d65b3926-signing-cabundle\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331529 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ba320f0-9564-46c3-8a4d-142fe14594d8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331558 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-service-ca\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331591 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns6t8\" (UniqueName: \"kubernetes.io/projected/2fcb6c86-c19d-4775-8a95-29e3d65b3926-kube-api-access-ns6t8\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331621 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd14138a-551f-4bbf-9934-b7555cb1d5e6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331638 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-trusted-ca\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331655 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg786\" (UniqueName: \"kubernetes.io/projected/a2bc1973-e22e-47e6-88bf-edab0829592e-kube-api-access-xg786\") pod \"control-plane-machine-set-operator-78cbb6b69f-zc5kp\" (UID: \"a2bc1973-e22e-47e6-88bf-edab0829592e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331691 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331720 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5e1e31e-b55a-4db6-8382-ede623b580d7-trusted-ca\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331742 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/58a0e454-50cb-4540-aeb6-b66435322029-tmpfs\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331777 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ba320f0-9564-46c3-8a4d-142fe14594d8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331794 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5ck4\" (UniqueName: \"kubernetes.io/projected/e4e50b25-8add-403e-bf95-b4825c548852-kube-api-access-n5ck4\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331821 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-bound-sa-token\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331863 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f1e06a-5f7f-413c-8682-62a523068334-config\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331879 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-default-certificate\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331895 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phsfk\" (UniqueName: \"kubernetes.io/projected/3bd11670-a667-45fe-a542-fdaabdb1900c-kube-api-access-phsfk\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331912 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59681aac-a82c-44a9-80bc-0870b8b51905-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331928 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb4a9851-d201-4a71-b26f-fff01b3c0e99-config\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331947 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd14138a-551f-4bbf-9934-b7555cb1d5e6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331963 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98478738-51a1-4daf-b1ca-34c21711fb40-config-volume\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.331978 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-srv-cert\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332010 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ba320f0-9564-46c3-8a4d-142fe14594d8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332029 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-metrics-certs\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332047 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j9t6\" (UniqueName: \"kubernetes.io/projected/e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5-kube-api-access-9j9t6\") pod \"migrator-59844c95c7-9mwwf\" (UID: \"e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332063 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e50b25-8add-403e-bf95-b4825c548852-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332080 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f36a8594-b24c-4df0-b946-3f979d311cfb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2w8gp\" (UID: \"f36a8594-b24c-4df0-b946-3f979d311cfb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332097 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332113 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-tls\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332128 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5n2w\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-kube-api-access-b5n2w\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332143 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb4a9851-d201-4a71-b26f-fff01b3c0e99-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332163 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-profile-collector-cert\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332177 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-stats-auth\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332197 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frzn6\" (UniqueName: \"kubernetes.io/projected/98478738-51a1-4daf-b1ca-34c21711fb40-kube-api-access-frzn6\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332213 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332229 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vcrq\" (UniqueName: \"kubernetes.io/projected/f36a8594-b24c-4df0-b946-3f979d311cfb-kube-api-access-7vcrq\") pod \"multus-admission-controller-857f4d67dd-2w8gp\" (UID: \"f36a8594-b24c-4df0-b946-3f979d311cfb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332242 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59681aac-a82c-44a9-80bc-0870b8b51905-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332255 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bd11670-a667-45fe-a542-fdaabdb1900c-service-ca-bundle\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332271 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pzxg\" (UniqueName: \"kubernetes.io/projected/70f428c8-e365-42aa-bced-3c9b721b6c90-kube-api-access-6pzxg\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332285 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zls6f\" (UniqueName: \"kubernetes.io/projected/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-kube-api-access-zls6f\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332309 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6fc4\" (UniqueName: \"kubernetes.io/projected/5e81408f-b0aa-4bde-be40-d132cd172ec9-kube-api-access-k6fc4\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332342 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98478738-51a1-4daf-b1ca-34c21711fb40-secret-volume\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332362 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5e1e31e-b55a-4db6-8382-ede623b580d7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332396 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-client\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332418 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/58a0e454-50cb-4540-aeb6-b66435322029-apiservice-cert\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.332435 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5e81408f-b0aa-4bde-be40-d132cd172ec9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.356803 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.377961 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.435292 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436066 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436357 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ba320f0-9564-46c3-8a4d-142fe14594d8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436391 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-bound-sa-token\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436416 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5ck4\" (UniqueName: \"kubernetes.io/projected/e4e50b25-8add-403e-bf95-b4825c548852-kube-api-access-n5ck4\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436502 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f1e06a-5f7f-413c-8682-62a523068334-config\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436522 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-default-certificate\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436542 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phsfk\" (UniqueName: \"kubernetes.io/projected/3bd11670-a667-45fe-a542-fdaabdb1900c-kube-api-access-phsfk\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436569 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59681aac-a82c-44a9-80bc-0870b8b51905-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436586 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb4a9851-d201-4a71-b26f-fff01b3c0e99-config\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436603 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd14138a-551f-4bbf-9934-b7555cb1d5e6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436620 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98478738-51a1-4daf-b1ca-34c21711fb40-config-volume\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436639 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-srv-cert\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.436661 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-socket-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.436707 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:42:59.936686472 +0000 UTC m=+136.707975451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.437923 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd14138a-551f-4bbf-9934-b7555cb1d5e6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.438640 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb4a9851-d201-4a71-b26f-fff01b3c0e99-config\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.438722 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98478738-51a1-4daf-b1ca-34c21711fb40-config-volume\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.439395 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f1e06a-5f7f-413c-8682-62a523068334-config\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.439522 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ba320f0-9564-46c3-8a4d-142fe14594d8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.439572 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-registration-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.439621 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-metrics-certs\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.439921 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e50b25-8add-403e-bf95-b4825c548852-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.440388 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j9t6\" (UniqueName: \"kubernetes.io/projected/e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5-kube-api-access-9j9t6\") pod \"migrator-59844c95c7-9mwwf\" (UID: \"e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.440518 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.440542 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f36a8594-b24c-4df0-b946-3f979d311cfb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2w8gp\" (UID: \"f36a8594-b24c-4df0-b946-3f979d311cfb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.440670 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4e50b25-8add-403e-bf95-b4825c548852-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442829 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-tls\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442855 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5n2w\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-kube-api-access-b5n2w\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442880 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb4a9851-d201-4a71-b26f-fff01b3c0e99-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442905 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-profile-collector-cert\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442921 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-stats-auth\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442951 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frzn6\" (UniqueName: \"kubernetes.io/projected/98478738-51a1-4daf-b1ca-34c21711fb40-kube-api-access-frzn6\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442974 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5cr9\" (UniqueName: \"kubernetes.io/projected/cdff737e-b5d7-45b1-ac32-f70d792f86fd-kube-api-access-c5cr9\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.442993 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443009 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vcrq\" (UniqueName: \"kubernetes.io/projected/f36a8594-b24c-4df0-b946-3f979d311cfb-kube-api-access-7vcrq\") pod \"multus-admission-controller-857f4d67dd-2w8gp\" (UID: \"f36a8594-b24c-4df0-b946-3f979d311cfb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443029 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/51b59561-4330-47c3-a6d2-25f5c76386fe-certs\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443045 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-967ck\" (UniqueName: \"kubernetes.io/projected/51b59561-4330-47c3-a6d2-25f5c76386fe-kube-api-access-967ck\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443063 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59681aac-a82c-44a9-80bc-0870b8b51905-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443078 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bd11670-a667-45fe-a542-fdaabdb1900c-service-ca-bundle\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443098 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pzxg\" (UniqueName: \"kubernetes.io/projected/70f428c8-e365-42aa-bced-3c9b721b6c90-kube-api-access-6pzxg\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443116 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zls6f\" (UniqueName: \"kubernetes.io/projected/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-kube-api-access-zls6f\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443148 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6fc4\" (UniqueName: \"kubernetes.io/projected/5e81408f-b0aa-4bde-be40-d132cd172ec9-kube-api-access-k6fc4\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443197 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cdff737e-b5d7-45b1-ac32-f70d792f86fd-metrics-tls\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443216 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98478738-51a1-4daf-b1ca-34c21711fb40-secret-volume\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443232 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgphd\" (UniqueName: \"kubernetes.io/projected/07be6ba5-ca8a-40e4-ac72-e0e013ab8d76-kube-api-access-hgphd\") pod \"ingress-canary-56v6b\" (UID: \"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76\") " pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443264 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhfg9\" (UniqueName: \"kubernetes.io/projected/188162b8-1946-4513-8dbf-a6c88b7af0b1-kube-api-access-hhfg9\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443295 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5e1e31e-b55a-4db6-8382-ede623b580d7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443326 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/58a0e454-50cb-4540-aeb6-b66435322029-apiservice-cert\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443354 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5e81408f-b0aa-4bde-be40-d132cd172ec9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443370 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-client\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443390 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f28t2\" (UniqueName: \"kubernetes.io/projected/11f1e06a-5f7f-413c-8682-62a523068334-kube-api-access-f28t2\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443406 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb4a9851-d201-4a71-b26f-fff01b3c0e99-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443421 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-csi-data-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443436 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5e1e31e-b55a-4db6-8382-ede623b580d7-metrics-tls\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443462 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/58a0e454-50cb-4540-aeb6-b66435322029-webhook-cert\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443504 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccz8g\" (UniqueName: \"kubernetes.io/projected/58a0e454-50cb-4540-aeb6-b66435322029-kube-api-access-ccz8g\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443524 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2bc1973-e22e-47e6-88bf-edab0829592e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zc5kp\" (UID: \"a2bc1973-e22e-47e6-88bf-edab0829592e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443554 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59681aac-a82c-44a9-80bc-0870b8b51905-config\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443592 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e50b25-8add-403e-bf95-b4825c548852-proxy-tls\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443620 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8hcs\" (UniqueName: \"kubernetes.io/projected/c5e1e31e-b55a-4db6-8382-ede623b580d7-kube-api-access-l8hcs\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443639 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjr84\" (UniqueName: \"kubernetes.io/projected/60bf2317-c2a3-4703-8e92-1c8d7c25a857-kube-api-access-mjr84\") pod \"package-server-manager-789f6589d5-jb6vd\" (UID: \"60bf2317-c2a3-4703-8e92-1c8d7c25a857\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443662 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443681 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp4x4\" (UniqueName: \"kubernetes.io/projected/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-kube-api-access-xp4x4\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443720 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-ca\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443752 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-certificates\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443776 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f1e06a-5f7f-413c-8682-62a523068334-serving-cert\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443791 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2fcb6c86-c19d-4775-8a95-29e3d65b3926-signing-key\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443808 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-serving-cert\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443826 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443851 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txwb6\" (UniqueName: \"kubernetes.io/projected/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-kube-api-access-txwb6\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443867 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-config\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443905 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/60bf2317-c2a3-4703-8e92-1c8d7c25a857-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jb6vd\" (UID: \"60bf2317-c2a3-4703-8e92-1c8d7c25a857\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443930 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5e81408f-b0aa-4bde-be40-d132cd172ec9-srv-cert\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443948 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2fcb6c86-c19d-4775-8a95-29e3d65b3926-signing-cabundle\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443964 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ba320f0-9564-46c3-8a4d-142fe14594d8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.443980 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07be6ba5-ca8a-40e4-ac72-e0e013ab8d76-cert\") pod \"ingress-canary-56v6b\" (UID: \"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76\") " pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444012 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-service-ca\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444041 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns6t8\" (UniqueName: \"kubernetes.io/projected/2fcb6c86-c19d-4775-8a95-29e3d65b3926-kube-api-access-ns6t8\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444059 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd14138a-551f-4bbf-9934-b7555cb1d5e6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444076 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-trusted-ca\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444092 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/51b59561-4330-47c3-a6d2-25f5c76386fe-node-bootstrap-token\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444120 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg786\" (UniqueName: \"kubernetes.io/projected/a2bc1973-e22e-47e6-88bf-edab0829592e-kube-api-access-xg786\") pod \"control-plane-machine-set-operator-78cbb6b69f-zc5kp\" (UID: \"a2bc1973-e22e-47e6-88bf-edab0829592e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444139 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-plugins-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444157 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444173 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdff737e-b5d7-45b1-ac32-f70d792f86fd-config-volume\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444190 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-mountpoint-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444226 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5e1e31e-b55a-4db6-8382-ede623b580d7-trusted-ca\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.444276 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/58a0e454-50cb-4540-aeb6-b66435322029-tmpfs\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.449109 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2fcb6c86-c19d-4775-8a95-29e3d65b3926-signing-cabundle\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.449119 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59681aac-a82c-44a9-80bc-0870b8b51905-config\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.449628 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ba320f0-9564-46c3-8a4d-142fe14594d8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.451068 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-ca\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.452057 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-certificates\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.454906 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-service-ca\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.457818 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:42:59.957802156 +0000 UTC m=+136.729091134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.457864 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.459359 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd14138a-551f-4bbf-9934-b7555cb1d5e6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.463787 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-trusted-ca\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.464114 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.464714 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-config\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.468576 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-tls\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.468848 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e4e50b25-8add-403e-bf95-b4825c548852-proxy-tls\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.471567 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/60bf2317-c2a3-4703-8e92-1c8d7c25a857-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jb6vd\" (UID: \"60bf2317-c2a3-4703-8e92-1c8d7c25a857\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.474202 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.476763 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98478738-51a1-4daf-b1ca-34c21711fb40-secret-volume\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.477129 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/58a0e454-50cb-4540-aeb6-b66435322029-tmpfs\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.477193 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a2bc1973-e22e-47e6-88bf-edab0829592e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zc5kp\" (UID: \"a2bc1973-e22e-47e6-88bf-edab0829592e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.479597 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-metrics-certs\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.479901 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-profile-collector-cert\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.481097 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bd11670-a667-45fe-a542-fdaabdb1900c-service-ca-bundle\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.481188 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f36a8594-b24c-4df0-b946-3f979d311cfb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2w8gp\" (UID: \"f36a8594-b24c-4df0-b946-3f979d311cfb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.481310 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-srv-cert\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.481389 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2fcb6c86-c19d-4775-8a95-29e3d65b3926-signing-key\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.481684 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f1e06a-5f7f-413c-8682-62a523068334-serving-cert\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.481987 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ba320f0-9564-46c3-8a4d-142fe14594d8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.485184 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-stats-auth\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.485652 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3bd11670-a667-45fe-a542-fdaabdb1900c-default-certificate\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.486528 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5e81408f-b0aa-4bde-be40-d132cd172ec9-srv-cert\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.487388 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.487736 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5e1e31e-b55a-4db6-8382-ede623b580d7-trusted-ca\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.489107 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c5e1e31e-b55a-4db6-8382-ede623b580d7-metrics-tls\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.490104 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-etcd-client\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.491270 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/58a0e454-50cb-4540-aeb6-b66435322029-apiservice-cert\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.493181 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59681aac-a82c-44a9-80bc-0870b8b51905-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.493495 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5e81408f-b0aa-4bde-be40-d132cd172ec9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.496186 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5ck4\" (UniqueName: \"kubernetes.io/projected/e4e50b25-8add-403e-bf95-b4825c548852-kube-api-access-n5ck4\") pod \"machine-config-controller-84d6567774-4d6pq\" (UID: \"e4e50b25-8add-403e-bf95-b4825c548852\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.503518 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/58a0e454-50cb-4540-aeb6-b66435322029-webhook-cert\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.506705 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-serving-cert\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.507827 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-bound-sa-token\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.507997 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fb4a9851-d201-4a71-b26f-fff01b3c0e99-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.508019 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1ba320f0-9564-46c3-8a4d-142fe14594d8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-w5kcl\" (UID: \"1ba320f0-9564-46c3-8a4d-142fe14594d8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.512553 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.527457 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/59681aac-a82c-44a9-80bc-0870b8b51905-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nzrh4\" (UID: \"59681aac-a82c-44a9-80bc-0870b8b51905\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.539457 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.545212 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.545671 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgphd\" (UniqueName: \"kubernetes.io/projected/07be6ba5-ca8a-40e4-ac72-e0e013ab8d76-kube-api-access-hgphd\") pod \"ingress-canary-56v6b\" (UID: \"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76\") " pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.545707 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhfg9\" (UniqueName: \"kubernetes.io/projected/188162b8-1946-4513-8dbf-a6c88b7af0b1-kube-api-access-hhfg9\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.545747 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-csi-data-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.545848 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07be6ba5-ca8a-40e4-ac72-e0e013ab8d76-cert\") pod \"ingress-canary-56v6b\" (UID: \"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76\") " pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.546012 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.546164 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.046148848 +0000 UTC m=+136.817437826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547213 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/51b59561-4330-47c3-a6d2-25f5c76386fe-node-bootstrap-token\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547243 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-plugins-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547264 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdff737e-b5d7-45b1-ac32-f70d792f86fd-config-volume\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547284 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-mountpoint-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547347 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-socket-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547385 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-registration-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547452 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5cr9\" (UniqueName: \"kubernetes.io/projected/cdff737e-b5d7-45b1-ac32-f70d792f86fd-kube-api-access-c5cr9\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547488 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/51b59561-4330-47c3-a6d2-25f5c76386fe-certs\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547505 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-967ck\" (UniqueName: \"kubernetes.io/projected/51b59561-4330-47c3-a6d2-25f5c76386fe-kube-api-access-967ck\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547545 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cdff737e-b5d7-45b1-ac32-f70d792f86fd-metrics-tls\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.547677 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-mountpoint-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.546272 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-csi-data-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.548171 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-socket-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.548185 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-registration-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.548210 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/188162b8-1946-4513-8dbf-a6c88b7af0b1-plugins-dir\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.548781 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cdff737e-b5d7-45b1-ac32-f70d792f86fd-config-volume\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.548925 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phsfk\" (UniqueName: \"kubernetes.io/projected/3bd11670-a667-45fe-a542-fdaabdb1900c-kube-api-access-phsfk\") pod \"router-default-5444994796-q28nt\" (UID: \"3bd11670-a667-45fe-a542-fdaabdb1900c\") " pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.555687 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.564988 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j9t6\" (UniqueName: \"kubernetes.io/projected/e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5-kube-api-access-9j9t6\") pod \"migrator-59844c95c7-9mwwf\" (UID: \"e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.565386 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/07be6ba5-ca8a-40e4-ac72-e0e013ab8d76-cert\") pod \"ingress-canary-56v6b\" (UID: \"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76\") " pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.565802 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/cdff737e-b5d7-45b1-ac32-f70d792f86fd-metrics-tls\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.569155 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/51b59561-4330-47c3-a6d2-25f5c76386fe-node-bootstrap-token\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.576950 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/51b59561-4330-47c3-a6d2-25f5c76386fe-certs\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.580765 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.599067 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccz8g\" (UniqueName: \"kubernetes.io/projected/58a0e454-50cb-4540-aeb6-b66435322029-kube-api-access-ccz8g\") pod \"packageserver-d55dfcdfc-tq22m\" (UID: \"58a0e454-50cb-4540-aeb6-b66435322029\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.613569 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wbkxj"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.625020 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5n2w\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-kube-api-access-b5n2w\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.634184 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.637336 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.638395 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fb4a9851-d201-4a71-b26f-fff01b3c0e99-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wqqkp\" (UID: \"fb4a9851-d201-4a71-b26f-fff01b3c0e99\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.648803 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.649197 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.14918445 +0000 UTC m=+136.920473428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.657455 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg786\" (UniqueName: \"kubernetes.io/projected/a2bc1973-e22e-47e6-88bf-edab0829592e-kube-api-access-xg786\") pod \"control-plane-machine-set-operator-78cbb6b69f-zc5kp\" (UID: \"a2bc1973-e22e-47e6-88bf-edab0829592e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.726309 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns6t8\" (UniqueName: \"kubernetes.io/projected/2fcb6c86-c19d-4775-8a95-29e3d65b3926-kube-api-access-ns6t8\") pod \"service-ca-9c57cc56f-vks6z\" (UID: \"2fcb6c86-c19d-4775-8a95-29e3d65b3926\") " pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.727787 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjr84\" (UniqueName: \"kubernetes.io/projected/60bf2317-c2a3-4703-8e92-1c8d7c25a857-kube-api-access-mjr84\") pod \"package-server-manager-789f6589d5-jb6vd\" (UID: \"60bf2317-c2a3-4703-8e92-1c8d7c25a857\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.728083 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-hr8jm"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.736101 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8hcs\" (UniqueName: \"kubernetes.io/projected/c5e1e31e-b55a-4db6-8382-ede623b580d7-kube-api-access-l8hcs\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.745651 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp4x4\" (UniqueName: \"kubernetes.io/projected/9f42e82b-55f3-4b00-8040-f5f020dc8f1e-kube-api-access-xp4x4\") pod \"etcd-operator-b45778765-b8qcp\" (UID: \"9f42e82b-55f3-4b00-8040-f5f020dc8f1e\") " pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.749526 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.750422 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.750986 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.250968449 +0000 UTC m=+137.022257427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.751101 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-n62tv"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.763767 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txwb6\" (UniqueName: \"kubernetes.io/projected/e0f38e73-7d25-48fc-80ce-8b727bb5a1b5-kube-api-access-txwb6\") pod \"catalog-operator-68c6474976-qtgnm\" (UID: \"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.786880 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.796876 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.804824 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vcrq\" (UniqueName: \"kubernetes.io/projected/f36a8594-b24c-4df0-b946-3f979d311cfb-kube-api-access-7vcrq\") pod \"multus-admission-controller-857f4d67dd-2w8gp\" (UID: \"f36a8594-b24c-4df0-b946-3f979d311cfb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.806609 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.841620 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.854767 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.857259 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.357154699 +0000 UTC m=+137.128443677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.862317 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.869446 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.882585 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frzn6\" (UniqueName: \"kubernetes.io/projected/98478738-51a1-4daf-b1ca-34c21711fb40-kube-api-access-frzn6\") pod \"collect-profiles-29526150-tswn4\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.887074 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c5e1e31e-b55a-4db6-8382-ede623b580d7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lzsfn\" (UID: \"c5e1e31e-b55a-4db6-8382-ede623b580d7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.890670 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.891223 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.895770 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.898499 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.898775 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pzxg\" (UniqueName: \"kubernetes.io/projected/70f428c8-e365-42aa-bced-3c9b721b6c90-kube-api-access-6pzxg\") pod \"marketplace-operator-79b997595-5snjq\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.904264 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zls6f\" (UniqueName: \"kubernetes.io/projected/985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae-kube-api-access-zls6f\") pod \"openshift-controller-manager-operator-756b6f6bc6-8nk5c\" (UID: \"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.913357 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6fc4\" (UniqueName: \"kubernetes.io/projected/5e81408f-b0aa-4bde-be40-d132cd172ec9-kube-api-access-k6fc4\") pod \"olm-operator-6b444d44fb-w64p6\" (UID: \"5e81408f-b0aa-4bde-be40-d132cd172ec9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.919246 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f28t2\" (UniqueName: \"kubernetes.io/projected/11f1e06a-5f7f-413c-8682-62a523068334-kube-api-access-f28t2\") pod \"service-ca-operator-777779d784-gblsk\" (UID: \"11f1e06a-5f7f-413c-8682-62a523068334\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.921247 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vt7z5"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.931079 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgphd\" (UniqueName: \"kubernetes.io/projected/07be6ba5-ca8a-40e4-ac72-e0e013ab8d76-kube-api-access-hgphd\") pod \"ingress-canary-56v6b\" (UID: \"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76\") " pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.947551 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhfg9\" (UniqueName: \"kubernetes.io/projected/188162b8-1946-4513-8dbf-a6c88b7af0b1-kube-api-access-hhfg9\") pod \"csi-hostpathplugin-7sf2k\" (UID: \"188162b8-1946-4513-8dbf-a6c88b7af0b1\") " pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.957564 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.957965 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.457934943 +0000 UTC m=+137.229223921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.958115 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:42:59 crc kubenswrapper[4492]: E0220 06:42:59.959312 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.459295221 +0000 UTC m=+137.230584199 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.963186 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5cr9\" (UniqueName: \"kubernetes.io/projected/cdff737e-b5d7-45b1-ac32-f70d792f86fd-kube-api-access-c5cr9\") pod \"dns-default-bmc25\" (UID: \"cdff737e-b5d7-45b1-ac32-f70d792f86fd\") " pod="openshift-dns/dns-default-bmc25" Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.988868 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p8dlg"] Feb 20 06:42:59 crc kubenswrapper[4492]: I0220 06:42:59.996194 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-967ck\" (UniqueName: \"kubernetes.io/projected/51b59561-4330-47c3-a6d2-25f5c76386fe-kube-api-access-967ck\") pod \"machine-config-server-psfnh\" (UID: \"51b59561-4330-47c3-a6d2-25f5c76386fe\") " pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:42:59.998629 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-gtnl8"] Feb 20 06:43:00 crc kubenswrapper[4492]: W0220 06:43:00.012959 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7923524_5488_4b05_a52f_7b198e0ba4ad.slice/crio-51e026dac477d374219b38e9147a710c635c0d087a32a6408e2dc92ebc4b4278 WatchSource:0}: Error finding container 51e026dac477d374219b38e9147a710c635c0d087a32a6408e2dc92ebc4b4278: Status 404 returned error can't find the container with id 51e026dac477d374219b38e9147a710c635c0d087a32a6408e2dc92ebc4b4278 Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.063353 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.065554 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.56552332 +0000 UTC m=+137.336812298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.065620 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.066114 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.566096753 +0000 UTC m=+137.337385731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.090791 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.099584 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" event={"ID":"010dc2a5-404a-47ed-881e-5cb6b5f6bae6","Type":"ContainerStarted","Data":"adf392f89f06e42040ada9ac106132bfa8ce53a33ae9fd2dcbe576145897b046"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.102553 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.117233 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-q28nt" event={"ID":"3bd11670-a667-45fe-a542-fdaabdb1900c","Type":"ContainerStarted","Data":"cd9537ec27cc6be26cc6f1ec4b5e8f8f770578418fd6fb2b3202635828983263"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.120446 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.124080 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.134544 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.134816 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" event={"ID":"327efe9d-4122-45d6-ad6b-62e41ae73a74","Type":"ContainerStarted","Data":"bfccb3eaeb6cfb8774abb6f9992f500c41ba9951cdaa0e60bc5f7e16a758b3b1"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.142175 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" event={"ID":"4c3dd185-faf8-4150-81b6-ef352c48148b","Type":"ContainerStarted","Data":"2833ea30aaba2004ea0bd7b1213778f736bc1d519e13a677d7c10ea0d7b2f93f"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.142213 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" event={"ID":"4c3dd185-faf8-4150-81b6-ef352c48148b","Type":"ContainerStarted","Data":"cc86486f95df67118d8742ec56b3327b375d7550203203e6097cabf8f161d222"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.145467 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wbkxj" event={"ID":"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb","Type":"ContainerStarted","Data":"01b2c9e52346030596d67613c35a98b9e1f9b177acf86f7ff60dc52d66d18135"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.150839 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.164678 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" event={"ID":"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7","Type":"ContainerStarted","Data":"2d54467a97c14729247bf16736ba09e1fc6e5a28b06821d1e4cd2278b90820a5"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.166810 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.167225 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.667211239 +0000 UTC m=+137.438500217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.169453 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-n62tv" event={"ID":"ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4","Type":"ContainerStarted","Data":"f8a374eea9d57eeb1381240200a4491873e70e3358fedaf3ea4179d037121ea9"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.170609 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" event={"ID":"471752d9-2f13-410f-831c-4785c2813fba","Type":"ContainerStarted","Data":"3e8654ee7133f82aa5da067ed2d4740e8be823752c274f2d163b9b96d0a2055a"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.174595 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.176238 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" event={"ID":"8f9c533a-3aa4-4413-914d-6d24f234c092","Type":"ContainerStarted","Data":"e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.176292 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" event={"ID":"8f9c533a-3aa4-4413-914d-6d24f234c092","Type":"ContainerStarted","Data":"f8f98e3bb00b3330ecc2d6b21ed8cc693311c4df7e4118178e78e24e75f35741"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.178283 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.189746 4492 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-v7ggr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.189924 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" podUID="8f9c533a-3aa4-4413-914d-6d24f234c092" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.201155 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bmc25" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.201972 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-56v6b" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.202320 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" event={"ID":"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3","Type":"ContainerStarted","Data":"8e1419ba2eb50804e085056f5c1bc9c76a604561e03e02387bbfc98d42f6aebb"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.203402 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" event={"ID":"c7923524-5488-4b05-a52f-7b198e0ba4ad","Type":"ContainerStarted","Data":"51e026dac477d374219b38e9147a710c635c0d087a32a6408e2dc92ebc4b4278"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.213868 4492 generic.go:334] "Generic (PLEG): container finished" podID="5559ffd4-7a7c-4603-b80e-0ea3159da609" containerID="46ca77255104444deec35169c763f3791ef252fd665e09303835f4580ea69ac1" exitCode=0 Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.215024 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" event={"ID":"5559ffd4-7a7c-4603-b80e-0ea3159da609","Type":"ContainerDied","Data":"46ca77255104444deec35169c763f3791ef252fd665e09303835f4580ea69ac1"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.215047 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" event={"ID":"5559ffd4-7a7c-4603-b80e-0ea3159da609","Type":"ContainerStarted","Data":"4107df66efc311bef55805ae4c2beb96db67f026b6b0dc952922f27770338a20"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.217028 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.217978 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" event={"ID":"123421da-365d-42ef-a5ef-1a129ec42490","Type":"ContainerStarted","Data":"94e42869b816eeb83d17e9aa4fcd8eb0ce36b301dbe1bbb4458ed1ab7c5432c3"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.225385 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" event={"ID":"de123619-8818-4fea-a9d0-75e498b66cca","Type":"ContainerStarted","Data":"949752f79feb85c74e8d382751356278b81f85126af8ec3184e83439ca834604"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.227167 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" event={"ID":"6fc4232f-8bca-43c5-a01e-02d2b58f43c9","Type":"ContainerStarted","Data":"4c50a4f9cec87e5190a83495525932e40f2d370bc8767ac7f2a2723b1821a681"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.228075 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.230607 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-psfnh" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.268101 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" event={"ID":"3be5a61d-eaaa-4eb7-bdc3-50d67493a117","Type":"ContainerStarted","Data":"be2b1460643f8924c6dab133695aa61a9f79ab3a06f94f91fb064e20a4eb133c"} Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.269550 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.769464223 +0000 UTC m=+137.540753201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.270209 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.278058 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" event={"ID":"80f87acb-f24f-4f79-9677-227c929e0497","Type":"ContainerStarted","Data":"7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.278328 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" event={"ID":"80f87acb-f24f-4f79-9677-227c929e0497","Type":"ContainerStarted","Data":"986edab851de5f00df0a850fdcf4fce712f5ba91fe01b7de1075f974ca9e8517"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.279060 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.285442 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.305434 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.310715 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" event={"ID":"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2","Type":"ContainerStarted","Data":"01928d14935792af121214c358d68590eb95cd833dc2a0d4a0b56a4f887cb416"} Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.375584 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.376515 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.876500598 +0000 UTC m=+137.647789566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.482935 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.485304 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:00.985279033 +0000 UTC m=+137.756568031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.586020 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.586558 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.086540415 +0000 UTC m=+137.857829394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.687611 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.688263 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.188249884 +0000 UTC m=+137.959538862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.735715 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" podStartSLOduration=119.735687973 podStartE2EDuration="1m59.735687973s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:00.734920615 +0000 UTC m=+137.506209592" watchObservedRunningTime="2026-02-20 06:43:00.735687973 +0000 UTC m=+137.506976951" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.773465 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.794806 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.795099 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.798324 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.798715 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.298699212 +0000 UTC m=+138.069988190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.809812 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" podStartSLOduration=119.809798206 podStartE2EDuration="1m59.809798206s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:00.808614492 +0000 UTC m=+137.579903470" watchObservedRunningTime="2026-02-20 06:43:00.809798206 +0000 UTC m=+137.581087184" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.832525 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-b8qcp"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.833925 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.844002 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.844032 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4"] Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.869446 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5clmz" podStartSLOduration=119.869423063 podStartE2EDuration="1m59.869423063s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:00.867903043 +0000 UTC m=+137.639192021" watchObservedRunningTime="2026-02-20 06:43:00.869423063 +0000 UTC m=+137.640712041" Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.900136 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:00 crc kubenswrapper[4492]: E0220 06:43:00.900446 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.400432796 +0000 UTC m=+138.171721774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:00 crc kubenswrapper[4492]: I0220 06:43:00.914705 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.002980 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.003545 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.503361516 +0000 UTC m=+138.274650495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.003575 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.004033 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.504026002 +0000 UTC m=+138.275314980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.006181 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-vks6z"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.048976 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-sjzvg" podStartSLOduration=120.048947359 podStartE2EDuration="2m0.048947359s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:01.04785143 +0000 UTC m=+137.819140408" watchObservedRunningTime="2026-02-20 06:43:01.048947359 +0000 UTC m=+137.820236338" Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.109052 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.109550 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.609534373 +0000 UTC m=+138.380823350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: W0220 06:43:01.204074 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58a0e454_50cb_4540_aeb6_b66435322029.slice/crio-6d2c45d6dbccb6580372e05f7e6832c1e78d49f66dad7e65c294fcb867d2ea46 WatchSource:0}: Error finding container 6d2c45d6dbccb6580372e05f7e6832c1e78d49f66dad7e65c294fcb867d2ea46: Status 404 returned error can't find the container with id 6d2c45d6dbccb6580372e05f7e6832c1e78d49f66dad7e65c294fcb867d2ea46 Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.210387 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.210858 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.710843695 +0000 UTC m=+138.482132673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.256627 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gblsk"] Feb 20 06:43:01 crc kubenswrapper[4492]: W0220 06:43:01.282510 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fcb6c86_c19d_4775_8a95_29e3d65b3926.slice/crio-2f5f5aa42ce21b406f922df9f83b0d66f2e3e95988fd038f8cae8c4fe5e38122 WatchSource:0}: Error finding container 2f5f5aa42ce21b406f922df9f83b0d66f2e3e95988fd038f8cae8c4fe5e38122: Status 404 returned error can't find the container with id 2f5f5aa42ce21b406f922df9f83b0d66f2e3e95988fd038f8cae8c4fe5e38122 Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.311946 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.312784 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.812765264 +0000 UTC m=+138.584054243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.399961 4492 generic.go:334] "Generic (PLEG): container finished" podID="6fc4232f-8bca-43c5-a01e-02d2b58f43c9" containerID="c322d3126a0a06ebfb280b4086302ac94c7a5039da340857896b6c6d622d1ef6" exitCode=0 Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.400257 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" event={"ID":"6fc4232f-8bca-43c5-a01e-02d2b58f43c9","Type":"ContainerDied","Data":"c322d3126a0a06ebfb280b4086302ac94c7a5039da340857896b6c6d622d1ef6"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.416086 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.417238 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:01.917225116 +0000 UTC m=+138.688514094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.422688 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" event={"ID":"471752d9-2f13-410f-831c-4785c2813fba","Type":"ContainerStarted","Data":"7de6c71b16ed80bcc0ef3a978a12ecc36a92a43587a4cfe00702c246598c20ad"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.443851 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2w8gp"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.467249 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.474801 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.506812 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" event={"ID":"1ba320f0-9564-46c3-8a4d-142fe14594d8","Type":"ContainerStarted","Data":"da60b4e81ac153395d88fbdc0ebd56172f75e8beadafa7237307b38fbc08e977"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.522960 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.524106 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.024090709 +0000 UTC m=+138.795379676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.532217 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" podStartSLOduration=120.532204427 podStartE2EDuration="2m0.532204427s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:01.531705835 +0000 UTC m=+138.302994812" watchObservedRunningTime="2026-02-20 06:43:01.532204427 +0000 UTC m=+138.303493405" Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.544830 4492 generic.go:334] "Generic (PLEG): container finished" podID="b5cd6268-e4fa-47ae-9cac-d03da9e84fe7" containerID="e5fc8066e1ebd09658c8a2a64bee6d731b34b995dd202abd9a0c2beb9abfc5af" exitCode=0 Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.544900 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" event={"ID":"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7","Type":"ContainerDied","Data":"e5fc8066e1ebd09658c8a2a64bee6d731b34b995dd202abd9a0c2beb9abfc5af"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.548619 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.570830 4492 csr.go:261] certificate signing request csr-27mpq is approved, waiting to be issued Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.577829 4492 csr.go:257] certificate signing request csr-27mpq is issued Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.589172 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" event={"ID":"123421da-365d-42ef-a5ef-1a129ec42490","Type":"ContainerStarted","Data":"d6484e507912d5d896ec0d96cc7820067bfdf62b675af6cc23a89ff2ba5bfa86"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.626826 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.627765 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.127748596 +0000 UTC m=+138.899037573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.635056 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-q28nt" event={"ID":"3bd11670-a667-45fe-a542-fdaabdb1900c","Type":"ContainerStarted","Data":"db1736b04c306096782901d367eefa5fb83a42042a816aa0d5109fb42925079f"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.669198 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" event={"ID":"58a0e454-50cb-4540-aeb6-b66435322029","Type":"ContainerStarted","Data":"6d2c45d6dbccb6580372e05f7e6832c1e78d49f66dad7e65c294fcb867d2ea46"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.682670 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" event={"ID":"e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5","Type":"ContainerStarted","Data":"d873d18659aba8aefc08a0f131d2d063effa224a5473b44db4da43624c28580b"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.692371 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5snjq"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.696003 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" event={"ID":"a2bc1973-e22e-47e6-88bf-edab0829592e","Type":"ContainerStarted","Data":"868cc459739809369a4de1c37bf6cffeca46ee0df320827a41e6118272af9196"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.706717 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" event={"ID":"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5","Type":"ContainerStarted","Data":"09cc4e25e2bd0e9d33b915f20f3d84356eeffdb53089ded9b4e6b8863a7e724c"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.728044 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.728160 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.228132912 +0000 UTC m=+138.999421890 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.728545 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.729896 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.229877516 +0000 UTC m=+139.001166494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.763388 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn"] Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.763947 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-n62tv" event={"ID":"ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4","Type":"ContainerStarted","Data":"ab72c1edfadb53cd2fc61bc08edbfdc5a58ff2b009ef8d4f8f6cd1900f41d946"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.765325 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-n62tv" Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.776108 4492 patch_prober.go:28] interesting pod/downloads-7954f5f757-n62tv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.776153 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-n62tv" podUID="ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.780271 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" event={"ID":"2fcb6c86-c19d-4775-8a95-29e3d65b3926","Type":"ContainerStarted","Data":"2f5f5aa42ce21b406f922df9f83b0d66f2e3e95988fd038f8cae8c4fe5e38122"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.785914 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" event={"ID":"6720f917-62d7-49d3-9bfb-ce7f9cfa89a2","Type":"ContainerStarted","Data":"b87187a4619e6908152067aa133fd42481eb5eb5ffd6cacd19618fbdb648948e"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.786668 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.791573 4492 patch_prober.go:28] interesting pod/console-operator-58897d9998-gtnl8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.791611 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" podUID="6720f917-62d7-49d3-9bfb-ce7f9cfa89a2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.808372 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" event={"ID":"fb4a9851-d201-4a71-b26f-fff01b3c0e99","Type":"ContainerStarted","Data":"fc40f1a551295aac6d77b4c5d1bc55518d0d259320abf0d1d04871df7cb12b6e"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.822653 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" event={"ID":"60bf2317-c2a3-4703-8e92-1c8d7c25a857","Type":"ContainerStarted","Data":"dfa391e1a8688d7f614eebabfa6c0ce4cfea56cd8fc150670cc1e36e12e6dc04"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.834803 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.835089 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.335072755 +0000 UTC m=+139.106361723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.835181 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.835526 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.335519278 +0000 UTC m=+139.106808257 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.838032 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" event={"ID":"59681aac-a82c-44a9-80bc-0870b8b51905","Type":"ContainerStarted","Data":"cc68cbd46bc1f7c22f237cd06be72f38f3bce4df03f9892279762f13fa3da384"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.863883 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" event={"ID":"e4e50b25-8add-403e-bf95-b4825c548852","Type":"ContainerStarted","Data":"f165d8dfdc2ebbe6eef2e0e5b979d706d09ec6566f716858c13f3216dd00338d"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.879136 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wbkxj" event={"ID":"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb","Type":"ContainerStarted","Data":"914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.896457 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" event={"ID":"9f42e82b-55f3-4b00-8040-f5f020dc8f1e","Type":"ContainerStarted","Data":"bd9f3f89c55ffee19f92e6bad608f73e8ea30c1c2ec61d7928548b8fa18a9a87"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.901752 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" event={"ID":"de123619-8818-4fea-a9d0-75e498b66cca","Type":"ContainerStarted","Data":"22231fbd97d6a987488c0b014bd2ca80c9a95686164a13235576d49a2d3c61cc"} Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.937072 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:01 crc kubenswrapper[4492]: E0220 06:43:01.938525 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.438508512 +0000 UTC m=+139.209797491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:01 crc kubenswrapper[4492]: I0220 06:43:01.994246 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" event={"ID":"327efe9d-4122-45d6-ad6b-62e41ae73a74","Type":"ContainerStarted","Data":"30df58b08e432a26f09b88edf51d1df7f227321519f57e1eba95e7bf7b80b992"} Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.018151 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-56v6b"] Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.024272 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-mfr74" event={"ID":"3be5a61d-eaaa-4eb7-bdc3-50d67493a117","Type":"ContainerStarted","Data":"3205feae7893675623f1bcbd200054de890ca2ba537f2b343ec1c288572ea8c1"} Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.033069 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lrqc8" podStartSLOduration=121.033052293 podStartE2EDuration="2m1.033052293s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:02.030690845 +0000 UTC m=+138.801979824" watchObservedRunningTime="2026-02-20 06:43:02.033052293 +0000 UTC m=+138.804341261" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.047371 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.048371 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.548352298 +0000 UTC m=+139.319641276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.054828 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" event={"ID":"010dc2a5-404a-47ed-881e-5cb6b5f6bae6","Type":"ContainerStarted","Data":"0e63ec606cb826f3a448f23f0f74eec1bf2ad4f093a0afc9779fb0793f41789b"} Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.101781 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.102315 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-n62tv" podStartSLOduration=121.102294417 podStartE2EDuration="2m1.102294417s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:02.101771149 +0000 UTC m=+138.873060126" watchObservedRunningTime="2026-02-20 06:43:02.102294417 +0000 UTC m=+138.873583395" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.137313 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jldvk" podStartSLOduration=121.137293692 podStartE2EDuration="2m1.137293692s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:02.122951234 +0000 UTC m=+138.894240213" watchObservedRunningTime="2026-02-20 06:43:02.137293692 +0000 UTC m=+138.908582669" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.137721 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bmc25"] Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.148145 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.152185 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.652170298 +0000 UTC m=+139.423459265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.169380 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-q28nt" podStartSLOduration=121.169357215 podStartE2EDuration="2m1.169357215s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:02.164487421 +0000 UTC m=+138.935776399" watchObservedRunningTime="2026-02-20 06:43:02.169357215 +0000 UTC m=+138.940646193" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.185597 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" podStartSLOduration=121.1855783 podStartE2EDuration="2m1.1855783s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:02.184621793 +0000 UTC m=+138.955910772" watchObservedRunningTime="2026-02-20 06:43:02.1855783 +0000 UTC m=+138.956867278" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.241680 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wbkxj" podStartSLOduration=121.24166228 podStartE2EDuration="2m1.24166228s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:02.240730562 +0000 UTC m=+139.012019540" watchObservedRunningTime="2026-02-20 06:43:02.24166228 +0000 UTC m=+139.012951258" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.251873 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.252263 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.752249649 +0000 UTC m=+139.523538627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.317575 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7sf2k"] Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.320429 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-hr8jm" podStartSLOduration=121.320406473 podStartE2EDuration="2m1.320406473s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:02.314710369 +0000 UTC m=+139.085999348" watchObservedRunningTime="2026-02-20 06:43:02.320406473 +0000 UTC m=+139.091695451" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.352883 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.353398 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.853382519 +0000 UTC m=+139.624671487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.454368 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.454917 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:02.954905745 +0000 UTC m=+139.726194724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.555867 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.557245 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.557251 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.05722212 +0000 UTC m=+139.828511098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.569913 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:02 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:02 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:02 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.569957 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.578859 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-20 06:38:01 +0000 UTC, rotation deadline is 2026-12-03 20:33:23.30437719 +0000 UTC Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.578887 4492 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6877h50m20.725492703s for next certificate rotation Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.657634 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.658047 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.158024356 +0000 UTC m=+139.929313334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.767155 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.767834 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.267820872 +0000 UTC m=+140.039109850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.871817 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:02 crc kubenswrapper[4492]: E0220 06:43:02.872244 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.372230599 +0000 UTC m=+140.143519577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:02 crc kubenswrapper[4492]: I0220 06:43:02.978276 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.010204 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.510185886 +0000 UTC m=+140.281474865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.083087 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.083704 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.583689785 +0000 UTC m=+140.354978763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.088035 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" event={"ID":"1ba320f0-9564-46c3-8a4d-142fe14594d8","Type":"ContainerStarted","Data":"5084b852677555f2dfa50b68b00e1b2a2f71c57975332e87dd431b8b59e2f0de"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.193903 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.195002 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.694982576 +0000 UTC m=+140.466271554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.198602 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" event={"ID":"de123619-8818-4fea-a9d0-75e498b66cca","Type":"ContainerStarted","Data":"aadf8b8198f77470704a4509ba7cb61e5b86fc8a0a5c609ecda6c983f20a1b78"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.250719 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" event={"ID":"e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5","Type":"ContainerStarted","Data":"4d86be915db69503cb6e73a4c8419b0ed2254ab46f3e5959bbdf49516e3db0c8"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.252114 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-w5kcl" podStartSLOduration=122.252091654 podStartE2EDuration="2m2.252091654s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.136517159 +0000 UTC m=+139.907806137" watchObservedRunningTime="2026-02-20 06:43:03.252091654 +0000 UTC m=+140.023380631" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.252970 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-66mcq" podStartSLOduration=122.252965193 podStartE2EDuration="2m2.252965193s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.251817736 +0000 UTC m=+140.023106713" watchObservedRunningTime="2026-02-20 06:43:03.252965193 +0000 UTC m=+140.024254171" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.283754 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" podStartSLOduration=122.283733651 podStartE2EDuration="2m2.283733651s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.282772796 +0000 UTC m=+140.054061775" watchObservedRunningTime="2026-02-20 06:43:03.283733651 +0000 UTC m=+140.055022629" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.285497 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-56v6b" event={"ID":"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76","Type":"ContainerStarted","Data":"fba711be10f8837607a7060869e482c992f51d67facf1fc695349aad3d9d2f94"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.288419 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" event={"ID":"11f1e06a-5f7f-413c-8682-62a523068334","Type":"ContainerStarted","Data":"7ade27967c446a67993792370eaba91a64356b986c881276008e561eeecadc4a"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.288448 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" event={"ID":"11f1e06a-5f7f-413c-8682-62a523068334","Type":"ContainerStarted","Data":"d4e8ad1903f48f637a74b29a2271b0426dfd1509c3182b0f766f8e3d472904da"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.295731 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.296068 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.796056595 +0000 UTC m=+140.567345573 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.311674 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-psfnh" event={"ID":"51b59561-4330-47c3-a6d2-25f5c76386fe","Type":"ContainerStarted","Data":"f3dde7047fea609b90666817f9055b47bf34c0c40353659cdb68ae20d827306f"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.311707 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-psfnh" event={"ID":"51b59561-4330-47c3-a6d2-25f5c76386fe","Type":"ContainerStarted","Data":"f336efe93759ee1d7262f44b94b61af87319baa064ae7ccc67ed5d8214491bd4"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.326669 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" event={"ID":"188162b8-1946-4513-8dbf-a6c88b7af0b1","Type":"ContainerStarted","Data":"3d615261798297e53f34620dd3782cfd98cd55ca7d55e409788f73e324c59cf9"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.356850 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" event={"ID":"58a0e454-50cb-4540-aeb6-b66435322029","Type":"ContainerStarted","Data":"53361a703974894afc083b3c630d4f77ce5db4dad946ce2249046f8fc3da7d3c"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.357727 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.360572 4492 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tq22m container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:5443/healthz\": dial tcp 10.217.0.23:5443: connect: connection refused" start-of-body= Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.360643 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" podUID="58a0e454-50cb-4540-aeb6-b66435322029" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.23:5443/healthz\": dial tcp 10.217.0.23:5443: connect: connection refused" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.381052 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" event={"ID":"b5cd6268-e4fa-47ae-9cac-d03da9e84fe7","Type":"ContainerStarted","Data":"9edeb1a638fdda48a36141c9f210837e4375f148986bf387167dc58f7b3213de"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.381292 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.390852 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gblsk" podStartSLOduration=122.390837944 podStartE2EDuration="2m2.390837944s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.34445132 +0000 UTC m=+140.115740298" watchObservedRunningTime="2026-02-20 06:43:03.390837944 +0000 UTC m=+140.162126922" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.394107 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" event={"ID":"c5e1e31e-b55a-4db6-8382-ede623b580d7","Type":"ContainerStarted","Data":"4a1dd32ac89183f7e8d3e3e9ebbd301508fe9410868915000af9e2a2d19dd3fb"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.394141 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" event={"ID":"c5e1e31e-b55a-4db6-8382-ede623b580d7","Type":"ContainerStarted","Data":"dd40f6ed4363b00cf69d11b2708043cfe8787f2d6ccba805927d94683b39bb30"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.396364 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.397535 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.897520761 +0000 UTC m=+140.668809739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.397645 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" event={"ID":"70f428c8-e365-42aa-bced-3c9b721b6c90","Type":"ContainerStarted","Data":"4c9c16a85a042558187c01a69930886ac41c156ea570331bdb97de52df5982c2"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.403332 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" event={"ID":"98478738-51a1-4daf-b1ca-34c21711fb40","Type":"ContainerStarted","Data":"e2ab5cbad3959bf11211df32c69a73b7dd2930688741c048d3d4cef63651375c"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.405219 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" event={"ID":"2fcb6c86-c19d-4775-8a95-29e3d65b3926","Type":"ContainerStarted","Data":"a8aa5b01ce55eef22807ef1cc1e79854c45fad65990937a1eb02c16927d60aae"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.406211 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" event={"ID":"5e81408f-b0aa-4bde-be40-d132cd172ec9","Type":"ContainerStarted","Data":"b303a1d4e483c349d12e6e7e27d70e634ae85ebec40f475b245334a0c3b384ec"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.407084 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" event={"ID":"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3","Type":"ContainerStarted","Data":"15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.407641 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.408387 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" event={"ID":"f36a8594-b24c-4df0-b946-3f979d311cfb","Type":"ContainerStarted","Data":"f74a5504e159a34e525e443f4493b1c74b2c2758adac7ae98c92a0123d922c9e"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.415689 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" event={"ID":"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae","Type":"ContainerStarted","Data":"a3e54e1485b18d81833fbfd8692336c893d4cc31c154ddd356c0f796f900b48a"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.435112 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-psfnh" podStartSLOduration=6.435098765 podStartE2EDuration="6.435098765s" podCreationTimestamp="2026-02-20 06:42:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.392756797 +0000 UTC m=+140.164045775" watchObservedRunningTime="2026-02-20 06:43:03.435098765 +0000 UTC m=+140.206387743" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.435190 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" podStartSLOduration=122.435186491 podStartE2EDuration="2m2.435186491s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.434289597 +0000 UTC m=+140.205578576" watchObservedRunningTime="2026-02-20 06:43:03.435186491 +0000 UTC m=+140.206475469" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.435527 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" event={"ID":"5559ffd4-7a7c-4603-b80e-0ea3159da609","Type":"ContainerStarted","Data":"fa9fda6e786c673cba13d1aa2009af91dbe006760ff2159f658928c1a9d263c6"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.443059 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" event={"ID":"60bf2317-c2a3-4703-8e92-1c8d7c25a857","Type":"ContainerStarted","Data":"e77f1bcfe33d439417baef77abb5ae8d415709936c8238afdde5ab8aa9514ae6"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.448066 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" event={"ID":"e4e50b25-8add-403e-bf95-b4825c548852","Type":"ContainerStarted","Data":"5325e25a00f0a56d24787ac0b71a9c082993f15c62057cfbee547cce657b4652"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.497785 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.499306 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:03.999293669 +0000 UTC m=+140.770582637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.500495 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" event={"ID":"e0f38e73-7d25-48fc-80ce-8b727bb5a1b5","Type":"ContainerStarted","Data":"a3404f7dce951115773abb67408ec43444da24aee267d2b2fa731712411c6ae7"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.501240 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.502812 4492 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qtgnm container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.502868 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" podUID="e0f38e73-7d25-48fc-80ce-8b727bb5a1b5" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.521254 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" podStartSLOduration=122.521240653 podStartE2EDuration="2m2.521240653s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.520309455 +0000 UTC m=+140.291598424" watchObservedRunningTime="2026-02-20 06:43:03.521240653 +0000 UTC m=+140.292529631" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.521528 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" podStartSLOduration=122.521524148 podStartE2EDuration="2m2.521524148s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.473785241 +0000 UTC m=+140.245074219" watchObservedRunningTime="2026-02-20 06:43:03.521524148 +0000 UTC m=+140.292813127" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.541680 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" event={"ID":"fb4a9851-d201-4a71-b26f-fff01b3c0e99","Type":"ContainerStarted","Data":"7e21436ea15fe2341dd0147ba884e2703defea00ad2176f6783a8d86b4e1a16a"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.546704 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bmc25" event={"ID":"cdff737e-b5d7-45b1-ac32-f70d792f86fd","Type":"ContainerStarted","Data":"2de8c14ef411ed1eb36bf332ff2742f282340611679f042c3d20369719708657"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.560538 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-vks6z" podStartSLOduration=122.560525499 podStartE2EDuration="2m2.560525499s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.559631922 +0000 UTC m=+140.330920900" watchObservedRunningTime="2026-02-20 06:43:03.560525499 +0000 UTC m=+140.331814477" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.575651 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:03 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:03 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:03 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.575685 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.579222 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" event={"ID":"010dc2a5-404a-47ed-881e-5cb6b5f6bae6","Type":"ContainerStarted","Data":"0c79458c5bb6623366f4d8a0fc6c02d35e44c1a3a73c14bf59bbc848eb81188a"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.599379 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.599645 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.099621389 +0000 UTC m=+140.870910367 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.600449 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.601206 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.10119496 +0000 UTC m=+140.872483928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.605172 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" event={"ID":"c7923524-5488-4b05-a52f-7b198e0ba4ad","Type":"ContainerStarted","Data":"1510a48c4bb7a67e501846079e902443d067b423693f7624ffaa6732182d2a76"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.697999 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" podStartSLOduration=122.697977246 podStartE2EDuration="2m2.697977246s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:03.696606157 +0000 UTC m=+140.467895136" watchObservedRunningTime="2026-02-20 06:43:03.697977246 +0000 UTC m=+140.469266224" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.702077 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" event={"ID":"6fc4232f-8bca-43c5-a01e-02d2b58f43c9","Type":"ContainerStarted","Data":"7cd7b6663604d94de6f0b2ffb19d3603fb649624ea08068e8b350b980e5a7035"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.710253 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.712065 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.212048031 +0000 UTC m=+140.983337009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.771773 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" event={"ID":"a2bc1973-e22e-47e6-88bf-edab0829592e","Type":"ContainerStarted","Data":"dacf4e838e5f2dcd6931e135cacdc4b56dc022fccaf2745ea4139644a1ad9c42"} Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.774651 4492 patch_prober.go:28] interesting pod/downloads-7954f5f757-n62tv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.774708 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-n62tv" podUID="ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.813311 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.814108 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.314090438 +0000 UTC m=+141.085379416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.913894 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:03 crc kubenswrapper[4492]: E0220 06:43:03.915158 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.415142746 +0000 UTC m=+141.186431724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.929879 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:43:03 crc kubenswrapper[4492]: I0220 06:43:03.929913 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.018283 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.018893 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.518882958 +0000 UTC m=+141.290171936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.079735 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.120145 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.120737 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.620717994 +0000 UTC m=+141.392006972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.189915 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wqqkp" podStartSLOduration=123.189896457 podStartE2EDuration="2m3.189896457s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:04.151820063 +0000 UTC m=+140.923109030" watchObservedRunningTime="2026-02-20 06:43:04.189896457 +0000 UTC m=+140.961185435" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.223772 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.223908 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" podStartSLOduration=123.223891617 podStartE2EDuration="2m3.223891617s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:04.223071198 +0000 UTC m=+140.994360176" watchObservedRunningTime="2026-02-20 06:43:04.223891617 +0000 UTC m=+140.995180595" Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.224037 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.724025359 +0000 UTC m=+141.495314338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.224049 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zc5kp" podStartSLOduration=123.224045658 podStartE2EDuration="2m3.224045658s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:04.190609944 +0000 UTC m=+140.961898923" watchObservedRunningTime="2026-02-20 06:43:04.224045658 +0000 UTC m=+140.995334635" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.257231 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qh4bk" podStartSLOduration=123.257219737 podStartE2EDuration="2m3.257219737s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:04.254737651 +0000 UTC m=+141.026026629" watchObservedRunningTime="2026-02-20 06:43:04.257219737 +0000 UTC m=+141.028508715" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.325618 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.326017 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.826004668 +0000 UTC m=+141.597293645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.378654 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.427193 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.428372 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:04.928340017 +0000 UTC m=+141.699628996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.531358 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.531863 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.031831029 +0000 UTC m=+141.803120007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.562686 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:04 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:04 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:04 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.562959 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.632794 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.633223 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.133206117 +0000 UTC m=+141.904495084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.733913 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.734202 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.234152184 +0000 UTC m=+142.005441162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.734519 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.735010 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.235001687 +0000 UTC m=+142.006290666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.770867 4492 patch_prober.go:28] interesting pod/console-operator-58897d9998-gtnl8 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.770933 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" podUID="6720f917-62d7-49d3-9bfb-ce7f9cfa89a2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.782298 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" event={"ID":"9f42e82b-55f3-4b00-8040-f5f020dc8f1e","Type":"ContainerStarted","Data":"54cbf03b40b608b5dda48aede1ad99f896aff7b765c28f71c29d61580b544eea"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.787102 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bmc25" event={"ID":"cdff737e-b5d7-45b1-ac32-f70d792f86fd","Type":"ContainerStarted","Data":"fb9f18e965173dec3dbe01dee085e4b2ff5037d068798ca88d02f5d942429141"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.787174 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bmc25" event={"ID":"cdff737e-b5d7-45b1-ac32-f70d792f86fd","Type":"ContainerStarted","Data":"bb91fe7ff12d11f1d468a9fa197825badcb59b5ad4fbb040b2545ff39f4735ba"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.787820 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-bmc25" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.790155 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" event={"ID":"5559ffd4-7a7c-4603-b80e-0ea3159da609","Type":"ContainerStarted","Data":"aeed437df54b9bb1af3b86ac42c554402a4e0d82a198e09985a48a0b504b438f"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.791704 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" event={"ID":"59681aac-a82c-44a9-80bc-0870b8b51905","Type":"ContainerStarted","Data":"513d40a55d8e7eff10a902df31d49778eeee0d4e6ec764b42101873331ff8112"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.796956 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" event={"ID":"70f428c8-e365-42aa-bced-3c9b721b6c90","Type":"ContainerStarted","Data":"13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.797221 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.798680 4492 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5snjq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.798718 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" podUID="70f428c8-e365-42aa-bced-3c9b721b6c90" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.798843 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" event={"ID":"60bf2317-c2a3-4703-8e92-1c8d7c25a857","Type":"ContainerStarted","Data":"f1671fbb4eb379298f2df04be276cf7a8b714b8ee57109e2238edb64eb4a26dd"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.799099 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.813543 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" event={"ID":"985ed43b-f1f5-40ad-9c6a-ca2b6d2d8cae","Type":"ContainerStarted","Data":"9e16eafb758b2158b61378396fc99465e3151f27061e2578a19cd96fc084e81c"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.816177 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9mwwf" event={"ID":"e9bd8bde-9b9a-4a7a-b5b4-26217e8dc9f5","Type":"ContainerStarted","Data":"5e36c870c29f1cd173ae61be13068e579d5df9586c9e68522d81ec714f50d63a"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.821121 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" event={"ID":"98478738-51a1-4daf-b1ca-34c21711fb40","Type":"ContainerStarted","Data":"375b3d46d8a7f93aae5dd415c252610d14e35734e3c403d6bd9ea4011919316b"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.835649 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.836217 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.336192547 +0000 UTC m=+142.107481525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.839498 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" event={"ID":"e4e50b25-8add-403e-bf95-b4825c548852","Type":"ContainerStarted","Data":"824582022d3f088bcfc025ec45378193d08925fd458ad4ddeb2f385aeb803f99"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.841830 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" event={"ID":"5e81408f-b0aa-4bde-be40-d132cd172ec9","Type":"ContainerStarted","Data":"fc9085e2b9d42208305606605c3240bba21f333abf051218d1776352d80538c5"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.842647 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.843837 4492 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-w64p6 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.843891 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" podUID="5e81408f-b0aa-4bde-be40-d132cd172ec9" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.864518 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" event={"ID":"188162b8-1946-4513-8dbf-a6c88b7af0b1","Type":"ContainerStarted","Data":"05fb26574ff4415deac1ba34e3ffe6b06dda62f16b111afb73810dc4b6854c25"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.883959 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-b8qcp" podStartSLOduration=123.883943608 podStartE2EDuration="2m3.883943608s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:04.854187099 +0000 UTC m=+141.625476068" watchObservedRunningTime="2026-02-20 06:43:04.883943608 +0000 UTC m=+141.655232575" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.885492 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" event={"ID":"c5e1e31e-b55a-4db6-8382-ede623b580d7","Type":"ContainerStarted","Data":"b9b1d75e72b3808148264fc7c84ab32f1b81ee2b060e39cbc13ca4d8a6464947"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.909813 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" event={"ID":"c7923524-5488-4b05-a52f-7b198e0ba4ad","Type":"ContainerStarted","Data":"84f1e8c41176f69a1648eee5f7282cdb9b958b729f4bb342dbae217c376cef00"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.922594 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-56v6b" event={"ID":"07be6ba5-ca8a-40e4-ac72-e0e013ab8d76","Type":"ContainerStarted","Data":"7ff46cccc4db0d6719e6ae85baebc2455334bebbce71e327593fef5eda28be43"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.937216 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" event={"ID":"f36a8594-b24c-4df0-b946-3f979d311cfb","Type":"ContainerStarted","Data":"39329113d46cae959b6e1a187874da96e22997c852e950ffbd179bb89ee22245"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.937246 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" event={"ID":"f36a8594-b24c-4df0-b946-3f979d311cfb","Type":"ContainerStarted","Data":"3910fd62517d37cf906091fb8dbd5a60be2bc7d646f04129c021d1e64a59a9ec"} Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.940936 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:04 crc kubenswrapper[4492]: E0220 06:43:04.944757 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.444741999 +0000 UTC m=+142.216030977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.958420 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t7xh6" Feb 20 06:43:04 crc kubenswrapper[4492]: I0220 06:43:04.978945 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qtgnm" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.050632 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.050795 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.550762536 +0000 UTC m=+142.322051514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.055052 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.056757 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.556739311 +0000 UTC m=+142.328028288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.074561 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" podStartSLOduration=124.074534656 podStartE2EDuration="2m4.074534656s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.072392302 +0000 UTC m=+141.843681290" watchObservedRunningTime="2026-02-20 06:43:05.074534656 +0000 UTC m=+141.845823634" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.162931 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.163101 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.663073329 +0000 UTC m=+142.434362298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.163370 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.163860 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.663849255 +0000 UTC m=+142.435138233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.171549 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bmc25" podStartSLOduration=9.171530065 podStartE2EDuration="9.171530065s" podCreationTimestamp="2026-02-20 06:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.169677788 +0000 UTC m=+141.940966766" watchObservedRunningTime="2026-02-20 06:43:05.171530065 +0000 UTC m=+141.942819043" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.172503 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzrh4" podStartSLOduration=124.172498072 podStartE2EDuration="2m4.172498072s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.13430124 +0000 UTC m=+141.905590218" watchObservedRunningTime="2026-02-20 06:43:05.172498072 +0000 UTC m=+141.943787051" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.220816 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" podStartSLOduration=124.22079363 podStartE2EDuration="2m4.22079363s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.220636994 +0000 UTC m=+141.991925973" watchObservedRunningTime="2026-02-20 06:43:05.22079363 +0000 UTC m=+141.992082608" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.222039 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-8nk5c" podStartSLOduration=124.222030736 podStartE2EDuration="2m4.222030736s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.202722284 +0000 UTC m=+141.974011262" watchObservedRunningTime="2026-02-20 06:43:05.222030736 +0000 UTC m=+141.993319715" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.265552 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.265745 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.765719035 +0000 UTC m=+142.537008013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.265875 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.266295 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.766283471 +0000 UTC m=+142.537572448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.326732 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" podStartSLOduration=124.32670403 podStartE2EDuration="2m4.32670403s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.270331303 +0000 UTC m=+142.041620281" watchObservedRunningTime="2026-02-20 06:43:05.32670403 +0000 UTC m=+142.097992997" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.366956 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.367266 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.867232334 +0000 UTC m=+142.638521312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.368043 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.368489 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.868458549 +0000 UTC m=+142.639747527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.398797 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" podStartSLOduration=124.398783179 podStartE2EDuration="2m4.398783179s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.330442769 +0000 UTC m=+142.101731736" watchObservedRunningTime="2026-02-20 06:43:05.398783179 +0000 UTC m=+142.170072157" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.401786 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-4d6pq" podStartSLOduration=124.401755441 podStartE2EDuration="2m4.401755441s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.398333079 +0000 UTC m=+142.169622057" watchObservedRunningTime="2026-02-20 06:43:05.401755441 +0000 UTC m=+142.173044419" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.468924 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.469076 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.969045348 +0000 UTC m=+142.740334316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.469262 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.469566 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:05.969552315 +0000 UTC m=+142.740841293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.491053 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" podStartSLOduration=124.491034552 podStartE2EDuration="2m4.491034552s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.426749446 +0000 UTC m=+142.198038425" watchObservedRunningTime="2026-02-20 06:43:05.491034552 +0000 UTC m=+142.262323530" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.509903 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-gtnl8" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.536615 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lzsfn" podStartSLOduration=124.53660245 podStartE2EDuration="2m4.53660245s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.534733581 +0000 UTC m=+142.306022550" watchObservedRunningTime="2026-02-20 06:43:05.53660245 +0000 UTC m=+142.307891417" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.536894 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vt7z5" podStartSLOduration=124.536890494 podStartE2EDuration="2m4.536890494s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.492588244 +0000 UTC m=+142.263877233" watchObservedRunningTime="2026-02-20 06:43:05.536890494 +0000 UTC m=+142.308179472" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.571224 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:05 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:05 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:05 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.571306 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.571822 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.572031 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.072008854 +0000 UTC m=+142.843297831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.572187 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.572691 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.07268456 +0000 UTC m=+142.843973538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.572790 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-56v6b" podStartSLOduration=9.572772165 podStartE2EDuration="9.572772165s" podCreationTimestamp="2026-02-20 06:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.570778982 +0000 UTC m=+142.342067960" watchObservedRunningTime="2026-02-20 06:43:05.572772165 +0000 UTC m=+142.344061143" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.658786 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xfgjc" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.659110 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-2w8gp" podStartSLOduration=124.659087651 podStartE2EDuration="2m4.659087651s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:05.65680419 +0000 UTC m=+142.428093169" watchObservedRunningTime="2026-02-20 06:43:05.659087651 +0000 UTC m=+142.430376629" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.673825 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.674649 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.1746228 +0000 UTC m=+142.945911778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.775101 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.775610 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.275587022 +0000 UTC m=+143.046876000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.876951 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.877224 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.377170482 +0000 UTC m=+143.148459460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.877511 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.878078 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.378068527 +0000 UTC m=+143.149357506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.940310 4492 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tq22m container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.940389 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" podUID="58a0e454-50cb-4540-aeb6-b66435322029" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.23:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.943975 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" event={"ID":"188162b8-1946-4513-8dbf-a6c88b7af0b1","Type":"ContainerStarted","Data":"84111994894af19026ce18f91544006c02fbb201b9497a816b83d2d6622ff51b"} Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.944770 4492 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5snjq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.944822 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" podUID="70f428c8-e365-42aa-bced-3c9b721b6c90" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.961611 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9f62m"] Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.962496 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.974785 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.978174 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:05 crc kubenswrapper[4492]: E0220 06:43:05.978588 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.478574094 +0000 UTC m=+143.249863071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:05 crc kubenswrapper[4492]: I0220 06:43:05.990374 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9f62m"] Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.055419 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-w64p6" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.080296 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-utilities\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.080533 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-catalog-content\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.080718 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nfdb\" (UniqueName: \"kubernetes.io/projected/8c489688-3ccb-4ad8-b309-da0d10f50618-kube-api-access-7nfdb\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.081333 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.100303 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.600271858 +0000 UTC m=+143.371560836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.171824 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lpshv"] Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.173189 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lpshv"] Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.173289 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.182336 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.182599 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-catalog-content\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.182655 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nfdb\" (UniqueName: \"kubernetes.io/projected/8c489688-3ccb-4ad8-b309-da0d10f50618-kube-api-access-7nfdb\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.182788 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-utilities\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.183182 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-utilities\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.183230 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.683198457 +0000 UTC m=+143.454487435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.183555 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-catalog-content\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.189377 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.245310 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nfdb\" (UniqueName: \"kubernetes.io/projected/8c489688-3ccb-4ad8-b309-da0d10f50618-kube-api-access-7nfdb\") pod \"community-operators-9f62m\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.274337 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.283937 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-utilities\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.283992 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-catalog-content\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.284066 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.284125 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klrjk\" (UniqueName: \"kubernetes.io/projected/563da73d-0987-4fef-93f2-acf2e5755392-kube-api-access-klrjk\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.284487 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.784459428 +0000 UTC m=+143.555748396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.339920 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8kmzx"] Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.341376 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.373875 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8kmzx"] Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.386004 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.386363 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klrjk\" (UniqueName: \"kubernetes.io/projected/563da73d-0987-4fef-93f2-acf2e5755392-kube-api-access-klrjk\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.386398 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-utilities\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.386425 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-catalog-content\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.388006 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.887988331 +0000 UTC m=+143.659277310 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.388680 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-catalog-content\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.388965 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-utilities\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.437108 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klrjk\" (UniqueName: \"kubernetes.io/projected/563da73d-0987-4fef-93f2-acf2e5755392-kube-api-access-klrjk\") pod \"certified-operators-lpshv\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.487669 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-utilities\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.487788 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.487861 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-catalog-content\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.487885 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw9jq\" (UniqueName: \"kubernetes.io/projected/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-kube-api-access-sw9jq\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.488371 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:06.988344265 +0000 UTC m=+143.759633243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.494981 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.564217 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:06 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:06 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:06 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.564286 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.573967 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f6t4g"] Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.575041 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.588315 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.588506 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.088456399 +0000 UTC m=+143.859745377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.589158 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-catalog-content\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.589213 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw9jq\" (UniqueName: \"kubernetes.io/projected/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-kube-api-access-sw9jq\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.589245 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-utilities\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.589342 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.589719 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.089704123 +0000 UTC m=+143.860993102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.589750 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-catalog-content\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.589857 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-utilities\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.613203 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw9jq\" (UniqueName: \"kubernetes.io/projected/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-kube-api-access-sw9jq\") pod \"community-operators-8kmzx\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.667228 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f6t4g"] Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.673403 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.695288 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.695442 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.195396221 +0000 UTC m=+143.966685189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.695721 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.695817 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb48k\" (UniqueName: \"kubernetes.io/projected/7e865a3a-b443-4e73-a75c-24980c388b6f-kube-api-access-pb48k\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.695849 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-utilities\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.695911 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-catalog-content\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.696349 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.196340053 +0000 UTC m=+143.967629031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.798034 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.798444 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb48k\" (UniqueName: \"kubernetes.io/projected/7e865a3a-b443-4e73-a75c-24980c388b6f-kube-api-access-pb48k\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.798515 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-utilities\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.799098 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-utilities\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.799508 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.298677227 +0000 UTC m=+144.069966205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.799630 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-catalog-content\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.799903 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-catalog-content\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.822200 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb48k\" (UniqueName: \"kubernetes.io/projected/7e865a3a-b443-4e73-a75c-24980c388b6f-kube-api-access-pb48k\") pod \"certified-operators-f6t4g\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.888295 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.901570 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:06 crc kubenswrapper[4492]: E0220 06:43:06.901988 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.401971296 +0000 UTC m=+144.173260275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.944953 4492 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tq22m container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.945044 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" podUID="58a0e454-50cb-4540-aeb6-b66435322029" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.23:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 20 06:43:06 crc kubenswrapper[4492]: I0220 06:43:06.980458 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" event={"ID":"188162b8-1946-4513-8dbf-a6c88b7af0b1","Type":"ContainerStarted","Data":"0421abbed900abfc8e380fa81766d09d8dc2eaadf99f6ccc04175b3e3a876a6d"} Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.004052 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.004599 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.504567059 +0000 UTC m=+144.275856027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.105622 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.107002 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.606982481 +0000 UTC m=+144.378271459 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.207103 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.207255 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.707234989 +0000 UTC m=+144.478523956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.207315 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.207626 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.707616949 +0000 UTC m=+144.478905917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.301103 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9f62m"] Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.308920 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.309130 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.809085653 +0000 UTC m=+144.580374631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.309345 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.309753 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.809741992 +0000 UTC m=+144.581030971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: W0220 06:43:07.316166 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c489688_3ccb_4ad8_b309_da0d10f50618.slice/crio-6b47cb8104fda9aaf7fcad56b2e7c1caa7b56ad2d281bf40a38835eb50f4daa3 WatchSource:0}: Error finding container 6b47cb8104fda9aaf7fcad56b2e7c1caa7b56ad2d281bf40a38835eb50f4daa3: Status 404 returned error can't find the container with id 6b47cb8104fda9aaf7fcad56b2e7c1caa7b56ad2d281bf40a38835eb50f4daa3 Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.410884 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.411133 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.911102042 +0000 UTC m=+144.682391020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.411274 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.411667 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:07.911658563 +0000 UTC m=+144.682947541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.512444 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.512608 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.012577809 +0000 UTC m=+144.783866788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.512697 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.513078 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.013069167 +0000 UTC m=+144.784358144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.520965 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lpshv"] Feb 20 06:43:07 crc kubenswrapper[4492]: W0220 06:43:07.568236 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod563da73d_0987_4fef_93f2_acf2e5755392.slice/crio-4c144757d0342f75c034ad245564ad01479053efee212ac8503168b9afbc3e51 WatchSource:0}: Error finding container 4c144757d0342f75c034ad245564ad01479053efee212ac8503168b9afbc3e51: Status 404 returned error can't find the container with id 4c144757d0342f75c034ad245564ad01479053efee212ac8503168b9afbc3e51 Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.570482 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:07 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:07 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:07 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.570516 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.614908 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.615580 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.115561905 +0000 UTC m=+144.886850882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.717311 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.717702 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.217684352 +0000 UTC m=+144.988973330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.738133 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8kmzx"] Feb 20 06:43:07 crc kubenswrapper[4492]: W0220 06:43:07.752346 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b57cdd0_38bd_4ad1_b0c2_87015ee8362a.slice/crio-aac0a0b77ce3a8ca77cace50abb922448e9611fb9994480bad913de7d1fb24e5 WatchSource:0}: Error finding container aac0a0b77ce3a8ca77cace50abb922448e9611fb9994480bad913de7d1fb24e5: Status 404 returned error can't find the container with id aac0a0b77ce3a8ca77cace50abb922448e9611fb9994480bad913de7d1fb24e5 Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.797587 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f6t4g"] Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.818773 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.819205 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.319174107 +0000 UTC m=+145.090463084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.819324 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.819879 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.319855893 +0000 UTC m=+145.091144871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.920660 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:07 crc kubenswrapper[4492]: E0220 06:43:07.921244 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.421225631 +0000 UTC m=+145.192514599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.993089 4492 generic.go:334] "Generic (PLEG): container finished" podID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerID="c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5" exitCode=0 Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.993316 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9f62m" event={"ID":"8c489688-3ccb-4ad8-b309-da0d10f50618","Type":"ContainerDied","Data":"c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5"} Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.993540 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9f62m" event={"ID":"8c489688-3ccb-4ad8-b309-da0d10f50618","Type":"ContainerStarted","Data":"6b47cb8104fda9aaf7fcad56b2e7c1caa7b56ad2d281bf40a38835eb50f4daa3"} Feb 20 06:43:07 crc kubenswrapper[4492]: I0220 06:43:07.995423 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.004345 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" event={"ID":"188162b8-1946-4513-8dbf-a6c88b7af0b1","Type":"ContainerStarted","Data":"0407d318262d3fe01802cae6b3392d886586680d49110e892c041ac863e015ca"} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.007922 4492 generic.go:334] "Generic (PLEG): container finished" podID="563da73d-0987-4fef-93f2-acf2e5755392" containerID="39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c" exitCode=0 Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.008009 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lpshv" event={"ID":"563da73d-0987-4fef-93f2-acf2e5755392","Type":"ContainerDied","Data":"39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c"} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.008039 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lpshv" event={"ID":"563da73d-0987-4fef-93f2-acf2e5755392","Type":"ContainerStarted","Data":"4c144757d0342f75c034ad245564ad01479053efee212ac8503168b9afbc3e51"} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.012501 4492 generic.go:334] "Generic (PLEG): container finished" podID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerID="1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8" exitCode=0 Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.012675 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kmzx" event={"ID":"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a","Type":"ContainerDied","Data":"1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8"} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.012726 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kmzx" event={"ID":"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a","Type":"ContainerStarted","Data":"aac0a0b77ce3a8ca77cace50abb922448e9611fb9994480bad913de7d1fb24e5"} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.016987 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t4g" event={"ID":"7e865a3a-b443-4e73-a75c-24980c388b6f","Type":"ContainerStarted","Data":"10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87"} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.017020 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t4g" event={"ID":"7e865a3a-b443-4e73-a75c-24980c388b6f","Type":"ContainerStarted","Data":"3c0ce24fc48d9d7682539c147a7c1d4f3c6239664dce61365fb72b728ffb65a7"} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.024675 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.025096 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.525080419 +0000 UTC m=+145.296369397 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.041528 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7sf2k" podStartSLOduration=12.041510868 podStartE2EDuration="12.041510868s" podCreationTimestamp="2026-02-20 06:42:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:08.039261421 +0000 UTC m=+144.810550400" watchObservedRunningTime="2026-02-20 06:43:08.041510868 +0000 UTC m=+144.812799836" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.123904 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8prdt"] Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.125564 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.126184 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.126373 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.626323497 +0000 UTC m=+145.397612475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.126903 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.127224 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.627211083 +0000 UTC m=+145.398500061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.127900 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.133664 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prdt"] Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.227880 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.228046 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.728020041 +0000 UTC m=+145.499309020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.228284 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-utilities\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.228320 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-catalog-content\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.228380 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7xsj\" (UniqueName: \"kubernetes.io/projected/23a7cdee-085f-4474-93f1-3edaea844222-kube-api-access-z7xsj\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.228727 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.229175 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.729157048 +0000 UTC m=+145.500446027 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.329976 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.330083 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.830064683 +0000 UTC m=+145.601353671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.330315 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.330396 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-utilities\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.330424 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-catalog-content\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.330508 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7xsj\" (UniqueName: \"kubernetes.io/projected/23a7cdee-085f-4474-93f1-3edaea844222-kube-api-access-z7xsj\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.330644 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.830636492 +0000 UTC m=+145.601925470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.330923 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-utilities\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.331106 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-catalog-content\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.354412 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7xsj\" (UniqueName: \"kubernetes.io/projected/23a7cdee-085f-4474-93f1-3edaea844222-kube-api-access-z7xsj\") pod \"redhat-marketplace-8prdt\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.431608 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.431734 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.931702135 +0000 UTC m=+145.702991114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.431863 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.432274 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:08.932264217 +0000 UTC m=+145.703553195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.440597 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.446504 4492 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.531372 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tqsgt"] Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.532604 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.534162 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.534753 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:09.034727208 +0000 UTC m=+145.806016186 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.535162 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.536140 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-20 06:43:09.036128974 +0000 UTC m=+145.807417952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-t9jrw" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.542380 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqsgt"] Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.560754 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:08 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:08 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:08 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.560831 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.636783 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.637213 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.637248 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.637312 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-utilities\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.637336 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.637389 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.637411 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp6l7\" (UniqueName: \"kubernetes.io/projected/25b6185c-71ad-476a-9dc9-97a5beeec215-kube-api-access-kp6l7\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.637433 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-catalog-content\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: E0220 06:43:08.637583 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-20 06:43:09.137563595 +0000 UTC m=+145.908852572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.639021 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.645104 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.648717 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prdt"] Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.654071 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.656459 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.671050 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.677828 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.683649 4492 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-20T06:43:08.446526062Z","Handler":null,"Name":""} Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.686274 4492 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.686407 4492 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.742516 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.743127 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-utilities\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.743695 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp6l7\" (UniqueName: \"kubernetes.io/projected/25b6185c-71ad-476a-9dc9-97a5beeec215-kube-api-access-kp6l7\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.743736 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-utilities\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.743813 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-catalog-content\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.744490 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-catalog-content\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.749071 4492 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.749108 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.769142 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.771511 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp6l7\" (UniqueName: \"kubernetes.io/projected/25b6185c-71ad-476a-9dc9-97a5beeec215-kube-api-access-kp6l7\") pod \"redhat-marketplace-tqsgt\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.811263 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-t9jrw\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.845843 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.856609 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.857490 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.917927 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.918031 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.931633 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:43:08 crc kubenswrapper[4492]: I0220 06:43:08.946596 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.066788 4492 patch_prober.go:28] interesting pod/downloads-7954f5f757-n62tv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.067238 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-n62tv" podUID="ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.069600 4492 patch_prober.go:28] interesting pod/downloads-7954f5f757-n62tv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" start-of-body= Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.069656 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-n62tv" podUID="ef7d81f8-c6d7-43c2-acd4-5a62577f0bf4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.11:8080/\": dial tcp 10.217.0.11:8080: connect: connection refused" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.076123 4492 generic.go:334] "Generic (PLEG): container finished" podID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerID="10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87" exitCode=0 Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.076182 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t4g" event={"ID":"7e865a3a-b443-4e73-a75c-24980c388b6f","Type":"ContainerDied","Data":"10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87"} Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.097082 4492 generic.go:334] "Generic (PLEG): container finished" podID="23a7cdee-085f-4474-93f1-3edaea844222" containerID="39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1" exitCode=0 Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.097143 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prdt" event={"ID":"23a7cdee-085f-4474-93f1-3edaea844222","Type":"ContainerDied","Data":"39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1"} Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.097166 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prdt" event={"ID":"23a7cdee-085f-4474-93f1-3edaea844222","Type":"ContainerStarted","Data":"b80ad750ef6b914fbe1049d856e3e7f536c134d4b07057a50b63755544ce4a94"} Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.141246 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7v5fb"] Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.142641 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.150026 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.167091 4492 generic.go:334] "Generic (PLEG): container finished" podID="98478738-51a1-4daf-b1ca-34c21711fb40" containerID="375b3d46d8a7f93aae5dd415c252610d14e35734e3c403d6bd9ea4011919316b" exitCode=0 Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.167302 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" event={"ID":"98478738-51a1-4daf-b1ca-34c21711fb40","Type":"ContainerDied","Data":"375b3d46d8a7f93aae5dd415c252610d14e35734e3c403d6bd9ea4011919316b"} Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.168251 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.168281 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.180894 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7v5fb"] Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.196911 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-t8jr5" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.211178 4492 patch_prober.go:28] interesting pod/console-f9d7485db-wbkxj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.211558 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wbkxj" podUID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.255889 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-utilities\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.255927 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r9hn\" (UniqueName: \"kubernetes.io/projected/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-kube-api-access-6r9hn\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.256085 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-catalog-content\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.311745 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.311785 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.358725 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-catalog-content\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.358819 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-utilities\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.358840 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r9hn\" (UniqueName: \"kubernetes.io/projected/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-kube-api-access-6r9hn\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.361840 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-catalog-content\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.362399 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-utilities\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.397089 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r9hn\" (UniqueName: \"kubernetes.io/projected/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-kube-api-access-6r9hn\") pod \"redhat-operators-7v5fb\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.481087 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.528229 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jtpfb"] Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.529170 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.546011 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtpfb"] Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.557556 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.559520 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:09 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:09 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:09 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.559551 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.665308 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnxm7\" (UniqueName: \"kubernetes.io/projected/4ba7350b-1379-405e-918c-022f4b2fd1b5-kube-api-access-pnxm7\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.665542 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-catalog-content\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.665572 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-utilities\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.730836 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.731421 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqsgt"] Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.731448 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.732074 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.734010 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.735342 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.740880 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.771349 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnxm7\" (UniqueName: \"kubernetes.io/projected/4ba7350b-1379-405e-918c-022f4b2fd1b5-kube-api-access-pnxm7\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.771612 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-catalog-content\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.771739 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-utilities\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.772358 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-utilities\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.772405 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-catalog-content\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.803045 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnxm7\" (UniqueName: \"kubernetes.io/projected/4ba7350b-1379-405e-918c-022f4b2fd1b5-kube-api-access-pnxm7\") pod \"redhat-operators-jtpfb\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.865229 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tq22m" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.872511 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2c0165-e948-449a-b11d-a75f4799966d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.872864 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2c0165-e948-449a-b11d-a75f4799966d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.933784 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.975512 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2c0165-e948-449a-b11d-a75f4799966d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.975636 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2c0165-e948-449a-b11d-a75f4799966d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.976920 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2c0165-e948-449a-b11d-a75f4799966d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:09 crc kubenswrapper[4492]: I0220 06:43:09.987886 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t9jrw"] Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.017949 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2c0165-e948-449a-b11d-a75f4799966d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.085068 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.139056 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.204264 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7f278dbc473b47c37fd5967aa5b3b8190e2e3c1f312ae78ee690dd67a78cecb9"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.204330 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ee68ba4cea013d0858f576224266b5d86e2a253acab3e97c769cddca518e2456"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.287762 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"725db09ff3bc454a1d81708d8ff76c615721137ddd877b06b908af2ec8cc3d6c"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.288015 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"426bc766ddc88e6364ac2bfe06dc20e640abcd2827366e9170d92ae1976bba39"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.315840 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7v5fb"] Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.330038 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" event={"ID":"bd14138a-551f-4bbf-9934-b7555cb1d5e6","Type":"ContainerStarted","Data":"873768df4a6f995f258639052b26a08756246c7f435cbd1afc76f152e74559aa"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.365727 4492 generic.go:334] "Generic (PLEG): container finished" podID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerID="a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419" exitCode=0 Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.365793 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqsgt" event={"ID":"25b6185c-71ad-476a-9dc9-97a5beeec215","Type":"ContainerDied","Data":"a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.365818 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqsgt" event={"ID":"25b6185c-71ad-476a-9dc9-97a5beeec215","Type":"ContainerStarted","Data":"6109633b47100ddc239ba70f201bdcbe643a60744b5c0f89b96e53a0d51f53a8"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.378923 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e19ab730c111f08cccb266f8838a4d6301d43c22aeae212b99e6c447e73fbeeb"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.378968 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"24f138b0dae359307343228b99435f8990287f93bd9b3bf481077ac197384732"} Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.379305 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.559137 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:10 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:10 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:10 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.559203 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.619530 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jtpfb"] Feb 20 06:43:10 crc kubenswrapper[4492]: I0220 06:43:10.670272 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.033987 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.113341 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frzn6\" (UniqueName: \"kubernetes.io/projected/98478738-51a1-4daf-b1ca-34c21711fb40-kube-api-access-frzn6\") pod \"98478738-51a1-4daf-b1ca-34c21711fb40\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.113501 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98478738-51a1-4daf-b1ca-34c21711fb40-secret-volume\") pod \"98478738-51a1-4daf-b1ca-34c21711fb40\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.113633 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98478738-51a1-4daf-b1ca-34c21711fb40-config-volume\") pod \"98478738-51a1-4daf-b1ca-34c21711fb40\" (UID: \"98478738-51a1-4daf-b1ca-34c21711fb40\") " Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.120997 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98478738-51a1-4daf-b1ca-34c21711fb40-config-volume" (OuterVolumeSpecName: "config-volume") pod "98478738-51a1-4daf-b1ca-34c21711fb40" (UID: "98478738-51a1-4daf-b1ca-34c21711fb40"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.128531 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98478738-51a1-4daf-b1ca-34c21711fb40-kube-api-access-frzn6" (OuterVolumeSpecName: "kube-api-access-frzn6") pod "98478738-51a1-4daf-b1ca-34c21711fb40" (UID: "98478738-51a1-4daf-b1ca-34c21711fb40"). InnerVolumeSpecName "kube-api-access-frzn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.131086 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98478738-51a1-4daf-b1ca-34c21711fb40-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "98478738-51a1-4daf-b1ca-34c21711fb40" (UID: "98478738-51a1-4daf-b1ca-34c21711fb40"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.217735 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frzn6\" (UniqueName: \"kubernetes.io/projected/98478738-51a1-4daf-b1ca-34c21711fb40-kube-api-access-frzn6\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.217769 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/98478738-51a1-4daf-b1ca-34c21711fb40-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.217782 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/98478738-51a1-4daf-b1ca-34c21711fb40-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.415600 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" event={"ID":"98478738-51a1-4daf-b1ca-34c21711fb40","Type":"ContainerDied","Data":"e2ab5cbad3959bf11211df32c69a73b7dd2930688741c048d3d4cef63651375c"} Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.415638 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2ab5cbad3959bf11211df32c69a73b7dd2930688741c048d3d4cef63651375c" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.415692 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.428667 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9e2c0165-e948-449a-b11d-a75f4799966d","Type":"ContainerStarted","Data":"afdc027877d9aeb5e6870af06f4ce5a0b4a53f0a4b263396ae65760650128250"} Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.435930 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtpfb" event={"ID":"4ba7350b-1379-405e-918c-022f4b2fd1b5","Type":"ContainerStarted","Data":"ade2643743f493b7b578aeb32845f3233902f3fa8cc408038715a1fe32131a85"} Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.455126 4492 generic.go:334] "Generic (PLEG): container finished" podID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerID="eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b" exitCode=0 Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.455188 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7v5fb" event={"ID":"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139","Type":"ContainerDied","Data":"eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b"} Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.455213 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7v5fb" event={"ID":"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139","Type":"ContainerStarted","Data":"dfd4c907ad8803890d0e1df875eef3611251b457ef7ab7c8d5735395535def50"} Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.459934 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" event={"ID":"bd14138a-551f-4bbf-9934-b7555cb1d5e6","Type":"ContainerStarted","Data":"1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a"} Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.460008 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.511672 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" podStartSLOduration=130.511628307 podStartE2EDuration="2m10.511628307s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:11.511318402 +0000 UTC m=+148.282607369" watchObservedRunningTime="2026-02-20 06:43:11.511628307 +0000 UTC m=+148.282917275" Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.564248 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:11 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:11 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:11 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:11 crc kubenswrapper[4492]: I0220 06:43:11.564296 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.244656 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 06:43:12 crc kubenswrapper[4492]: E0220 06:43:12.245284 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98478738-51a1-4daf-b1ca-34c21711fb40" containerName="collect-profiles" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.245295 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="98478738-51a1-4daf-b1ca-34c21711fb40" containerName="collect-profiles" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.245414 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="98478738-51a1-4daf-b1ca-34c21711fb40" containerName="collect-profiles" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.246882 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.253785 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.256187 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.256426 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.341979 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f4c666b-afd3-44bc-b917-dbcaa182510d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.342066 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f4c666b-afd3-44bc-b917-dbcaa182510d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.444957 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f4c666b-afd3-44bc-b917-dbcaa182510d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.445142 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f4c666b-afd3-44bc-b917-dbcaa182510d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.445220 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f4c666b-afd3-44bc-b917-dbcaa182510d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.464686 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f4c666b-afd3-44bc-b917-dbcaa182510d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.482811 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9e2c0165-e948-449a-b11d-a75f4799966d","Type":"ContainerStarted","Data":"708dc9b0dd5096532ca94b5c697b410331fcb0a04ba3f105a28c07ede5f409f7"} Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.494913 4492 generic.go:334] "Generic (PLEG): container finished" podID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerID="4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8" exitCode=0 Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.496197 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtpfb" event={"ID":"4ba7350b-1379-405e-918c-022f4b2fd1b5","Type":"ContainerDied","Data":"4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8"} Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.532947 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.532930816 podStartE2EDuration="3.532930816s" podCreationTimestamp="2026-02-20 06:43:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:12.522585025 +0000 UTC m=+149.293874003" watchObservedRunningTime="2026-02-20 06:43:12.532930816 +0000 UTC m=+149.304219794" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.561924 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:12 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:12 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:12 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.561980 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:12 crc kubenswrapper[4492]: I0220 06:43:12.569688 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:13 crc kubenswrapper[4492]: I0220 06:43:13.517933 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 20 06:43:13 crc kubenswrapper[4492]: W0220 06:43:13.546407 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2f4c666b_afd3_44bc_b917_dbcaa182510d.slice/crio-9a9e51eeec913b9a8c20be802aa137bea038962ec2cfe62288f513f250118a78 WatchSource:0}: Error finding container 9a9e51eeec913b9a8c20be802aa137bea038962ec2cfe62288f513f250118a78: Status 404 returned error can't find the container with id 9a9e51eeec913b9a8c20be802aa137bea038962ec2cfe62288f513f250118a78 Feb 20 06:43:13 crc kubenswrapper[4492]: I0220 06:43:13.557196 4492 generic.go:334] "Generic (PLEG): container finished" podID="9e2c0165-e948-449a-b11d-a75f4799966d" containerID="708dc9b0dd5096532ca94b5c697b410331fcb0a04ba3f105a28c07ede5f409f7" exitCode=0 Feb 20 06:43:13 crc kubenswrapper[4492]: I0220 06:43:13.565405 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:13 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:13 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:13 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:13 crc kubenswrapper[4492]: I0220 06:43:13.565548 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:13 crc kubenswrapper[4492]: I0220 06:43:13.586942 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9e2c0165-e948-449a-b11d-a75f4799966d","Type":"ContainerDied","Data":"708dc9b0dd5096532ca94b5c697b410331fcb0a04ba3f105a28c07ede5f409f7"} Feb 20 06:43:14 crc kubenswrapper[4492]: I0220 06:43:14.562604 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:14 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:14 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:14 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:14 crc kubenswrapper[4492]: I0220 06:43:14.562869 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:14 crc kubenswrapper[4492]: I0220 06:43:14.677289 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2f4c666b-afd3-44bc-b917-dbcaa182510d","Type":"ContainerStarted","Data":"4041dcb33fad6ba372cf395b0693acb9cad6037c7159032ff3cacf8fc22c0f37"} Feb 20 06:43:14 crc kubenswrapper[4492]: I0220 06:43:14.677358 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2f4c666b-afd3-44bc-b917-dbcaa182510d","Type":"ContainerStarted","Data":"9a9e51eeec913b9a8c20be802aa137bea038962ec2cfe62288f513f250118a78"} Feb 20 06:43:14 crc kubenswrapper[4492]: I0220 06:43:14.695342 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.6953223729999998 podStartE2EDuration="2.695322373s" podCreationTimestamp="2026-02-20 06:43:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:14.690863284 +0000 UTC m=+151.462152262" watchObservedRunningTime="2026-02-20 06:43:14.695322373 +0000 UTC m=+151.466611351" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.204792 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bmc25" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.229853 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.342253 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2c0165-e948-449a-b11d-a75f4799966d-kubelet-dir\") pod \"9e2c0165-e948-449a-b11d-a75f4799966d\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.342867 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2c0165-e948-449a-b11d-a75f4799966d-kube-api-access\") pod \"9e2c0165-e948-449a-b11d-a75f4799966d\" (UID: \"9e2c0165-e948-449a-b11d-a75f4799966d\") " Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.344510 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e2c0165-e948-449a-b11d-a75f4799966d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e2c0165-e948-449a-b11d-a75f4799966d" (UID: "9e2c0165-e948-449a-b11d-a75f4799966d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.364799 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e2c0165-e948-449a-b11d-a75f4799966d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e2c0165-e948-449a-b11d-a75f4799966d" (UID: "9e2c0165-e948-449a-b11d-a75f4799966d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.444740 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e2c0165-e948-449a-b11d-a75f4799966d-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.444770 4492 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e2c0165-e948-449a-b11d-a75f4799966d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.560744 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:15 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:15 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:15 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.560806 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.688193 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2f4c666b-afd3-44bc-b917-dbcaa182510d","Type":"ContainerDied","Data":"4041dcb33fad6ba372cf395b0693acb9cad6037c7159032ff3cacf8fc22c0f37"} Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.688693 4492 generic.go:334] "Generic (PLEG): container finished" podID="2f4c666b-afd3-44bc-b917-dbcaa182510d" containerID="4041dcb33fad6ba372cf395b0693acb9cad6037c7159032ff3cacf8fc22c0f37" exitCode=0 Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.695643 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"9e2c0165-e948-449a-b11d-a75f4799966d","Type":"ContainerDied","Data":"afdc027877d9aeb5e6870af06f4ce5a0b4a53f0a4b263396ae65760650128250"} Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.695681 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 20 06:43:15 crc kubenswrapper[4492]: I0220 06:43:15.695686 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afdc027877d9aeb5e6870af06f4ce5a0b4a53f0a4b263396ae65760650128250" Feb 20 06:43:16 crc kubenswrapper[4492]: I0220 06:43:16.558895 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:16 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:16 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:16 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:16 crc kubenswrapper[4492]: I0220 06:43:16.558941 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.182523 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.285558 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f4c666b-afd3-44bc-b917-dbcaa182510d-kube-api-access\") pod \"2f4c666b-afd3-44bc-b917-dbcaa182510d\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.286271 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f4c666b-afd3-44bc-b917-dbcaa182510d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2f4c666b-afd3-44bc-b917-dbcaa182510d" (UID: "2f4c666b-afd3-44bc-b917-dbcaa182510d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.286301 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f4c666b-afd3-44bc-b917-dbcaa182510d-kubelet-dir\") pod \"2f4c666b-afd3-44bc-b917-dbcaa182510d\" (UID: \"2f4c666b-afd3-44bc-b917-dbcaa182510d\") " Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.290197 4492 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f4c666b-afd3-44bc-b917-dbcaa182510d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.291070 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4c666b-afd3-44bc-b917-dbcaa182510d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2f4c666b-afd3-44bc-b917-dbcaa182510d" (UID: "2f4c666b-afd3-44bc-b917-dbcaa182510d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.391745 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f4c666b-afd3-44bc-b917-dbcaa182510d-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.560159 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:17 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:17 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:17 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.560204 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.752140 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"2f4c666b-afd3-44bc-b917-dbcaa182510d","Type":"ContainerDied","Data":"9a9e51eeec913b9a8c20be802aa137bea038962ec2cfe62288f513f250118a78"} Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.752177 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a9e51eeec913b9a8c20be802aa137bea038962ec2cfe62288f513f250118a78" Feb 20 06:43:17 crc kubenswrapper[4492]: I0220 06:43:17.752229 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 20 06:43:18 crc kubenswrapper[4492]: I0220 06:43:18.558997 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:18 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:18 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:18 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:18 crc kubenswrapper[4492]: I0220 06:43:18.559232 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:19 crc kubenswrapper[4492]: I0220 06:43:19.098964 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-n62tv" Feb 20 06:43:19 crc kubenswrapper[4492]: I0220 06:43:19.168319 4492 patch_prober.go:28] interesting pod/console-f9d7485db-wbkxj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 20 06:43:19 crc kubenswrapper[4492]: I0220 06:43:19.168371 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wbkxj" podUID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 20 06:43:19 crc kubenswrapper[4492]: I0220 06:43:19.559397 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:19 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:19 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:19 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:19 crc kubenswrapper[4492]: I0220 06:43:19.559460 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:20 crc kubenswrapper[4492]: I0220 06:43:20.560135 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:20 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:20 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:20 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:20 crc kubenswrapper[4492]: I0220 06:43:20.560616 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:21 crc kubenswrapper[4492]: I0220 06:43:21.561199 4492 patch_prober.go:28] interesting pod/router-default-5444994796-q28nt container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 20 06:43:21 crc kubenswrapper[4492]: [-]has-synced failed: reason withheld Feb 20 06:43:21 crc kubenswrapper[4492]: [+]process-running ok Feb 20 06:43:21 crc kubenswrapper[4492]: healthz check failed Feb 20 06:43:21 crc kubenswrapper[4492]: I0220 06:43:21.561468 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-q28nt" podUID="3bd11670-a667-45fe-a542-fdaabdb1900c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 20 06:43:22 crc kubenswrapper[4492]: I0220 06:43:22.560308 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:43:22 crc kubenswrapper[4492]: I0220 06:43:22.562881 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-q28nt" Feb 20 06:43:22 crc kubenswrapper[4492]: I0220 06:43:22.794711 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:43:22 crc kubenswrapper[4492]: I0220 06:43:22.810266 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0825f068-2c78-44a0-8f24-679e4521f069-metrics-certs\") pod \"network-metrics-daemon-p9vbm\" (UID: \"0825f068-2c78-44a0-8f24-679e4521f069\") " pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:43:22 crc kubenswrapper[4492]: I0220 06:43:22.874082 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p9vbm" Feb 20 06:43:25 crc kubenswrapper[4492]: I0220 06:43:25.319440 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p9vbm"] Feb 20 06:43:25 crc kubenswrapper[4492]: I0220 06:43:25.831729 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" event={"ID":"0825f068-2c78-44a0-8f24-679e4521f069","Type":"ContainerStarted","Data":"62c840bdf5469165dc5649f47a5d54e27ea79973c4b55ebe1f7a051a6c4ffa88"} Feb 20 06:43:25 crc kubenswrapper[4492]: I0220 06:43:25.831977 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" event={"ID":"0825f068-2c78-44a0-8f24-679e4521f069","Type":"ContainerStarted","Data":"c02ec3fc15222ad6dfb217e5668f5f81e34090a91902f18da2c26f46127b035f"} Feb 20 06:43:26 crc kubenswrapper[4492]: I0220 06:43:26.840110 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p9vbm" event={"ID":"0825f068-2c78-44a0-8f24-679e4521f069","Type":"ContainerStarted","Data":"2c048b82a5177e058d28a9b6e9995c501c0491720ffb43c868011d35d56011ae"} Feb 20 06:43:26 crc kubenswrapper[4492]: I0220 06:43:26.858324 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-p9vbm" podStartSLOduration=145.858310049 podStartE2EDuration="2m25.858310049s" podCreationTimestamp="2026-02-20 06:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:26.85318895 +0000 UTC m=+163.624477928" watchObservedRunningTime="2026-02-20 06:43:26.858310049 +0000 UTC m=+163.629599016" Feb 20 06:43:28 crc kubenswrapper[4492]: I0220 06:43:28.954608 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:43:29 crc kubenswrapper[4492]: I0220 06:43:29.172136 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:43:29 crc kubenswrapper[4492]: I0220 06:43:29.176760 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.451032 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.451760 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pb48k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-f6t4g_openshift-marketplace(7e865a3a-b443-4e73-a75c-24980c388b6f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.452939 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-f6t4g" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.480944 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.481111 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-klrjk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lpshv_openshift-marketplace(563da73d-0987-4fef-93f2-acf2e5755392): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.484623 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lpshv" podUID="563da73d-0987-4fef-93f2-acf2e5755392" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.501364 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.501464 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6r9hn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-7v5fb_openshift-marketplace(2fcf3fe3-c5f4-40a4-8b92-06d408c9d139): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 06:43:37 crc kubenswrapper[4492]: E0220 06:43:37.503556 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-7v5fb" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" Feb 20 06:43:38 crc kubenswrapper[4492]: E0220 06:43:38.899304 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-7v5fb" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" Feb 20 06:43:38 crc kubenswrapper[4492]: E0220 06:43:38.899333 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lpshv" podUID="563da73d-0987-4fef-93f2-acf2e5755392" Feb 20 06:43:38 crc kubenswrapper[4492]: E0220 06:43:38.899373 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-f6t4g" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" Feb 20 06:43:38 crc kubenswrapper[4492]: E0220 06:43:38.987120 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 20 06:43:38 crc kubenswrapper[4492]: E0220 06:43:38.987305 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z7xsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8prdt_openshift-marketplace(23a7cdee-085f-4474-93f1-3edaea844222): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 06:43:38 crc kubenswrapper[4492]: E0220 06:43:38.988745 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-8prdt" podUID="23a7cdee-085f-4474-93f1-3edaea844222" Feb 20 06:43:39 crc kubenswrapper[4492]: E0220 06:43:39.008576 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 20 06:43:39 crc kubenswrapper[4492]: E0220 06:43:39.008846 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kp6l7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-tqsgt_openshift-marketplace(25b6185c-71ad-476a-9dc9-97a5beeec215): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 20 06:43:39 crc kubenswrapper[4492]: E0220 06:43:39.010122 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-tqsgt" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.311513 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.311580 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.875412 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jb6vd" Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.953246 4492 generic.go:334] "Generic (PLEG): container finished" podID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerID="1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a" exitCode=0 Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.953309 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kmzx" event={"ID":"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a","Type":"ContainerDied","Data":"1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a"} Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.957276 4492 generic.go:334] "Generic (PLEG): container finished" podID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerID="04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0" exitCode=0 Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.957325 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtpfb" event={"ID":"4ba7350b-1379-405e-918c-022f4b2fd1b5","Type":"ContainerDied","Data":"04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0"} Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.960728 4492 generic.go:334] "Generic (PLEG): container finished" podID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerID="96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2" exitCode=0 Feb 20 06:43:39 crc kubenswrapper[4492]: I0220 06:43:39.960782 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9f62m" event={"ID":"8c489688-3ccb-4ad8-b309-da0d10f50618","Type":"ContainerDied","Data":"96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2"} Feb 20 06:43:39 crc kubenswrapper[4492]: E0220 06:43:39.961890 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-tqsgt" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" Feb 20 06:43:39 crc kubenswrapper[4492]: E0220 06:43:39.963402 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8prdt" podUID="23a7cdee-085f-4474-93f1-3edaea844222" Feb 20 06:43:40 crc kubenswrapper[4492]: I0220 06:43:40.971283 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtpfb" event={"ID":"4ba7350b-1379-405e-918c-022f4b2fd1b5","Type":"ContainerStarted","Data":"356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f"} Feb 20 06:43:40 crc kubenswrapper[4492]: I0220 06:43:40.972938 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9f62m" event={"ID":"8c489688-3ccb-4ad8-b309-da0d10f50618","Type":"ContainerStarted","Data":"151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271"} Feb 20 06:43:40 crc kubenswrapper[4492]: I0220 06:43:40.974900 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kmzx" event={"ID":"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a","Type":"ContainerStarted","Data":"22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9"} Feb 20 06:43:40 crc kubenswrapper[4492]: I0220 06:43:40.988901 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jtpfb" podStartSLOduration=4.092221428 podStartE2EDuration="31.988886208s" podCreationTimestamp="2026-02-20 06:43:09 +0000 UTC" firstStartedPulling="2026-02-20 06:43:12.523695231 +0000 UTC m=+149.294984209" lastFinishedPulling="2026-02-20 06:43:40.420360011 +0000 UTC m=+177.191648989" observedRunningTime="2026-02-20 06:43:40.987051765 +0000 UTC m=+177.758340743" watchObservedRunningTime="2026-02-20 06:43:40.988886208 +0000 UTC m=+177.760175176" Feb 20 06:43:41 crc kubenswrapper[4492]: I0220 06:43:41.036598 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9f62m" podStartSLOduration=3.597987168 podStartE2EDuration="36.036583277s" podCreationTimestamp="2026-02-20 06:43:05 +0000 UTC" firstStartedPulling="2026-02-20 06:43:07.995103706 +0000 UTC m=+144.766392675" lastFinishedPulling="2026-02-20 06:43:40.433699805 +0000 UTC m=+177.204988784" observedRunningTime="2026-02-20 06:43:41.034148489 +0000 UTC m=+177.805437467" watchObservedRunningTime="2026-02-20 06:43:41.036583277 +0000 UTC m=+177.807872255" Feb 20 06:43:41 crc kubenswrapper[4492]: I0220 06:43:41.037298 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8kmzx" podStartSLOduration=2.602474006 podStartE2EDuration="35.037294077s" podCreationTimestamp="2026-02-20 06:43:06 +0000 UTC" firstStartedPulling="2026-02-20 06:43:08.014607948 +0000 UTC m=+144.785896926" lastFinishedPulling="2026-02-20 06:43:40.449428019 +0000 UTC m=+177.220716997" observedRunningTime="2026-02-20 06:43:41.013226239 +0000 UTC m=+177.784515218" watchObservedRunningTime="2026-02-20 06:43:41.037294077 +0000 UTC m=+177.808583055" Feb 20 06:43:46 crc kubenswrapper[4492]: I0220 06:43:46.275788 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:46 crc kubenswrapper[4492]: I0220 06:43:46.276469 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:46 crc kubenswrapper[4492]: I0220 06:43:46.355098 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:46 crc kubenswrapper[4492]: I0220 06:43:46.674789 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:46 crc kubenswrapper[4492]: I0220 06:43:46.674871 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:46 crc kubenswrapper[4492]: I0220 06:43:46.707081 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:47 crc kubenswrapper[4492]: I0220 06:43:47.066866 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:47 crc kubenswrapper[4492]: I0220 06:43:47.067665 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:43:47 crc kubenswrapper[4492]: I0220 06:43:47.547492 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8kmzx"] Feb 20 06:43:47 crc kubenswrapper[4492]: I0220 06:43:47.866921 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p8dlg"] Feb 20 06:43:48 crc kubenswrapper[4492]: I0220 06:43:48.676504 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.040763 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8kmzx" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="registry-server" containerID="cri-o://22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9" gracePeriod=2 Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.351556 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.396018 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw9jq\" (UniqueName: \"kubernetes.io/projected/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-kube-api-access-sw9jq\") pod \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.396187 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-catalog-content\") pod \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.396219 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-utilities\") pod \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\" (UID: \"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a\") " Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.397061 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-utilities" (OuterVolumeSpecName: "utilities") pod "4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" (UID: "4b57cdd0-38bd-4ad1-b0c2-87015ee8362a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.402163 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-kube-api-access-sw9jq" (OuterVolumeSpecName: "kube-api-access-sw9jq") pod "4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" (UID: "4b57cdd0-38bd-4ad1-b0c2-87015ee8362a"). InnerVolumeSpecName "kube-api-access-sw9jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.414599 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 06:43:49 crc kubenswrapper[4492]: E0220 06:43:49.414993 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="extract-utilities" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415020 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="extract-utilities" Feb 20 06:43:49 crc kubenswrapper[4492]: E0220 06:43:49.415040 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4c666b-afd3-44bc-b917-dbcaa182510d" containerName="pruner" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415048 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4c666b-afd3-44bc-b917-dbcaa182510d" containerName="pruner" Feb 20 06:43:49 crc kubenswrapper[4492]: E0220 06:43:49.415061 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="registry-server" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415068 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="registry-server" Feb 20 06:43:49 crc kubenswrapper[4492]: E0220 06:43:49.415083 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e2c0165-e948-449a-b11d-a75f4799966d" containerName="pruner" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415088 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e2c0165-e948-449a-b11d-a75f4799966d" containerName="pruner" Feb 20 06:43:49 crc kubenswrapper[4492]: E0220 06:43:49.415096 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="extract-content" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415102 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="extract-content" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415245 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4c666b-afd3-44bc-b917-dbcaa182510d" containerName="pruner" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415277 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e2c0165-e948-449a-b11d-a75f4799966d" containerName="pruner" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415288 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerName="registry-server" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.415943 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.423042 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.423147 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.428635 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.447155 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" (UID: "4b57cdd0-38bd-4ad1-b0c2-87015ee8362a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.497887 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.497964 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.498018 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw9jq\" (UniqueName: \"kubernetes.io/projected/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-kube-api-access-sw9jq\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.498042 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.498053 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.599228 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.599293 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.599372 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.614310 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.746207 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.934765 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.934956 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:49 crc kubenswrapper[4492]: I0220 06:43:49.966038 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.049037 4492 generic.go:334] "Generic (PLEG): container finished" podID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" containerID="22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9" exitCode=0 Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.049128 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kmzx" event={"ID":"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a","Type":"ContainerDied","Data":"22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9"} Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.049148 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8kmzx" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.049167 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8kmzx" event={"ID":"4b57cdd0-38bd-4ad1-b0c2-87015ee8362a","Type":"ContainerDied","Data":"aac0a0b77ce3a8ca77cace50abb922448e9611fb9994480bad913de7d1fb24e5"} Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.049190 4492 scope.go:117] "RemoveContainer" containerID="22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.080171 4492 scope.go:117] "RemoveContainer" containerID="1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.080341 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8kmzx"] Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.083515 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.083627 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8kmzx"] Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.106172 4492 scope.go:117] "RemoveContainer" containerID="1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.113264 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.122035 4492 scope.go:117] "RemoveContainer" containerID="22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9" Feb 20 06:43:50 crc kubenswrapper[4492]: E0220 06:43:50.122548 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9\": container with ID starting with 22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9 not found: ID does not exist" containerID="22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.122597 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9"} err="failed to get container status \"22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9\": rpc error: code = NotFound desc = could not find container \"22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9\": container with ID starting with 22a2fca1647e312b04f40a3038c05aeeb32879a9e0ff368dd430239e19241bf9 not found: ID does not exist" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.122660 4492 scope.go:117] "RemoveContainer" containerID="1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a" Feb 20 06:43:50 crc kubenswrapper[4492]: E0220 06:43:50.123017 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a\": container with ID starting with 1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a not found: ID does not exist" containerID="1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.123103 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a"} err="failed to get container status \"1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a\": rpc error: code = NotFound desc = could not find container \"1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a\": container with ID starting with 1d263f297f2b49218d9b25adb40e60a31db35fdbdae419ae89448e4a925e299a not found: ID does not exist" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.123178 4492 scope.go:117] "RemoveContainer" containerID="1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8" Feb 20 06:43:50 crc kubenswrapper[4492]: E0220 06:43:50.123558 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8\": container with ID starting with 1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8 not found: ID does not exist" containerID="1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8" Feb 20 06:43:50 crc kubenswrapper[4492]: I0220 06:43:50.123576 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8"} err="failed to get container status \"1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8\": rpc error: code = NotFound desc = could not find container \"1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8\": container with ID starting with 1c26f1371909b84428743f92bbdb4784af6c7ff02555035402ee389483cd9da8 not found: ID does not exist" Feb 20 06:43:50 crc kubenswrapper[4492]: W0220 06:43:50.126849 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5f4ffa8d_dace_4d21_9774_1ca614e5671e.slice/crio-5eededf8f1f6edf10fd1e069c74a84515bc2aa6f0c87a363b48ad142c5ee965f WatchSource:0}: Error finding container 5eededf8f1f6edf10fd1e069c74a84515bc2aa6f0c87a363b48ad142c5ee965f: Status 404 returned error can't find the container with id 5eededf8f1f6edf10fd1e069c74a84515bc2aa6f0c87a363b48ad142c5ee965f Feb 20 06:43:51 crc kubenswrapper[4492]: I0220 06:43:51.056846 4492 generic.go:334] "Generic (PLEG): container finished" podID="5f4ffa8d-dace-4d21-9774-1ca614e5671e" containerID="7e5239a9aa31984d42d6b94f04a77840a2666da3ffb019247998c91a39a47454" exitCode=0 Feb 20 06:43:51 crc kubenswrapper[4492]: I0220 06:43:51.056921 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5f4ffa8d-dace-4d21-9774-1ca614e5671e","Type":"ContainerDied","Data":"7e5239a9aa31984d42d6b94f04a77840a2666da3ffb019247998c91a39a47454"} Feb 20 06:43:51 crc kubenswrapper[4492]: I0220 06:43:51.056956 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5f4ffa8d-dace-4d21-9774-1ca614e5671e","Type":"ContainerStarted","Data":"5eededf8f1f6edf10fd1e069c74a84515bc2aa6f0c87a363b48ad142c5ee965f"} Feb 20 06:43:51 crc kubenswrapper[4492]: I0220 06:43:51.563329 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b57cdd0-38bd-4ad1-b0c2-87015ee8362a" path="/var/lib/kubelet/pods/4b57cdd0-38bd-4ad1-b0c2-87015ee8362a/volumes" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.069363 4492 generic.go:334] "Generic (PLEG): container finished" podID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerID="b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d" exitCode=0 Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.069449 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqsgt" event={"ID":"25b6185c-71ad-476a-9dc9-97a5beeec215","Type":"ContainerDied","Data":"b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d"} Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.274612 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.334488 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kube-api-access\") pod \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.334693 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kubelet-dir\") pod \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\" (UID: \"5f4ffa8d-dace-4d21-9774-1ca614e5671e\") " Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.334952 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5f4ffa8d-dace-4d21-9774-1ca614e5671e" (UID: "5f4ffa8d-dace-4d21-9774-1ca614e5671e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.336308 4492 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.341061 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5f4ffa8d-dace-4d21-9774-1ca614e5671e" (UID: "5f4ffa8d-dace-4d21-9774-1ca614e5671e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.346785 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtpfb"] Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.347068 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jtpfb" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="registry-server" containerID="cri-o://356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f" gracePeriod=2 Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.437693 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f4ffa8d-dace-4d21-9774-1ca614e5671e-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.627395 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.742156 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnxm7\" (UniqueName: \"kubernetes.io/projected/4ba7350b-1379-405e-918c-022f4b2fd1b5-kube-api-access-pnxm7\") pod \"4ba7350b-1379-405e-918c-022f4b2fd1b5\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.742287 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-catalog-content\") pod \"4ba7350b-1379-405e-918c-022f4b2fd1b5\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.742316 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-utilities\") pod \"4ba7350b-1379-405e-918c-022f4b2fd1b5\" (UID: \"4ba7350b-1379-405e-918c-022f4b2fd1b5\") " Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.743000 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-utilities" (OuterVolumeSpecName: "utilities") pod "4ba7350b-1379-405e-918c-022f4b2fd1b5" (UID: "4ba7350b-1379-405e-918c-022f4b2fd1b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.746468 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba7350b-1379-405e-918c-022f4b2fd1b5-kube-api-access-pnxm7" (OuterVolumeSpecName: "kube-api-access-pnxm7") pod "4ba7350b-1379-405e-918c-022f4b2fd1b5" (UID: "4ba7350b-1379-405e-918c-022f4b2fd1b5"). InnerVolumeSpecName "kube-api-access-pnxm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.837173 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ba7350b-1379-405e-918c-022f4b2fd1b5" (UID: "4ba7350b-1379-405e-918c-022f4b2fd1b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.844456 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.844576 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba7350b-1379-405e-918c-022f4b2fd1b5-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:52 crc kubenswrapper[4492]: I0220 06:43:52.844652 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnxm7\" (UniqueName: \"kubernetes.io/projected/4ba7350b-1379-405e-918c-022f4b2fd1b5-kube-api-access-pnxm7\") on node \"crc\" DevicePath \"\"" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.092089 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqsgt" event={"ID":"25b6185c-71ad-476a-9dc9-97a5beeec215","Type":"ContainerStarted","Data":"04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7"} Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.095071 4492 generic.go:334] "Generic (PLEG): container finished" podID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerID="356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f" exitCode=0 Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.095250 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jtpfb" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.096314 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtpfb" event={"ID":"4ba7350b-1379-405e-918c-022f4b2fd1b5","Type":"ContainerDied","Data":"356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f"} Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.096439 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jtpfb" event={"ID":"4ba7350b-1379-405e-918c-022f4b2fd1b5","Type":"ContainerDied","Data":"ade2643743f493b7b578aeb32845f3233902f3fa8cc408038715a1fe32131a85"} Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.096507 4492 scope.go:117] "RemoveContainer" containerID="356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.097949 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"5f4ffa8d-dace-4d21-9774-1ca614e5671e","Type":"ContainerDied","Data":"5eededf8f1f6edf10fd1e069c74a84515bc2aa6f0c87a363b48ad142c5ee965f"} Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.097969 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eededf8f1f6edf10fd1e069c74a84515bc2aa6f0c87a363b48ad142c5ee965f" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.097967 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.101340 4492 generic.go:334] "Generic (PLEG): container finished" podID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerID="555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930" exitCode=0 Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.101390 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t4g" event={"ID":"7e865a3a-b443-4e73-a75c-24980c388b6f","Type":"ContainerDied","Data":"555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930"} Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.114296 4492 scope.go:117] "RemoveContainer" containerID="04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.139984 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tqsgt" podStartSLOduration=2.954928295 podStartE2EDuration="45.139965653s" podCreationTimestamp="2026-02-20 06:43:08 +0000 UTC" firstStartedPulling="2026-02-20 06:43:10.3736218 +0000 UTC m=+147.144910777" lastFinishedPulling="2026-02-20 06:43:52.558659157 +0000 UTC m=+189.329948135" observedRunningTime="2026-02-20 06:43:53.1167107 +0000 UTC m=+189.887999678" watchObservedRunningTime="2026-02-20 06:43:53.139965653 +0000 UTC m=+189.911254631" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.145852 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jtpfb"] Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.147392 4492 scope.go:117] "RemoveContainer" containerID="4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.155230 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jtpfb"] Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.179365 4492 scope.go:117] "RemoveContainer" containerID="356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f" Feb 20 06:43:53 crc kubenswrapper[4492]: E0220 06:43:53.180756 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f\": container with ID starting with 356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f not found: ID does not exist" containerID="356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.180839 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f"} err="failed to get container status \"356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f\": rpc error: code = NotFound desc = could not find container \"356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f\": container with ID starting with 356673fcc950a8ba909e7c43ac729408c221e7cbc623c3c8f6fac0e2d3621a2f not found: ID does not exist" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.180926 4492 scope.go:117] "RemoveContainer" containerID="04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0" Feb 20 06:43:53 crc kubenswrapper[4492]: E0220 06:43:53.181216 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0\": container with ID starting with 04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0 not found: ID does not exist" containerID="04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.181285 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0"} err="failed to get container status \"04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0\": rpc error: code = NotFound desc = could not find container \"04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0\": container with ID starting with 04d662dfe48325658b99c020c8a59be05194deedd48c3f78cf9f939365efb8e0 not found: ID does not exist" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.181345 4492 scope.go:117] "RemoveContainer" containerID="4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8" Feb 20 06:43:53 crc kubenswrapper[4492]: E0220 06:43:53.184757 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8\": container with ID starting with 4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8 not found: ID does not exist" containerID="4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.184849 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8"} err="failed to get container status \"4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8\": rpc error: code = NotFound desc = could not find container \"4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8\": container with ID starting with 4d570adaeb14c1963b9063caa9d40d44c16f600a3e0d70ad6b3c62fda8c9b1b8 not found: ID does not exist" Feb 20 06:43:53 crc kubenswrapper[4492]: I0220 06:43:53.565027 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" path="/var/lib/kubelet/pods/4ba7350b-1379-405e-918c-022f4b2fd1b5/volumes" Feb 20 06:43:54 crc kubenswrapper[4492]: I0220 06:43:54.111240 4492 generic.go:334] "Generic (PLEG): container finished" podID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerID="7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6" exitCode=0 Feb 20 06:43:54 crc kubenswrapper[4492]: I0220 06:43:54.111306 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7v5fb" event={"ID":"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139","Type":"ContainerDied","Data":"7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6"} Feb 20 06:43:54 crc kubenswrapper[4492]: I0220 06:43:54.114090 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t4g" event={"ID":"7e865a3a-b443-4e73-a75c-24980c388b6f","Type":"ContainerStarted","Data":"812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7"} Feb 20 06:43:54 crc kubenswrapper[4492]: I0220 06:43:54.152574 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f6t4g" podStartSLOduration=2.584888911 podStartE2EDuration="48.152548391s" podCreationTimestamp="2026-02-20 06:43:06 +0000 UTC" firstStartedPulling="2026-02-20 06:43:08.019094478 +0000 UTC m=+144.790383457" lastFinishedPulling="2026-02-20 06:43:53.586753959 +0000 UTC m=+190.358042937" observedRunningTime="2026-02-20 06:43:54.152112827 +0000 UTC m=+190.923401805" watchObservedRunningTime="2026-02-20 06:43:54.152548391 +0000 UTC m=+190.923837369" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.125632 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7v5fb" event={"ID":"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139","Type":"ContainerStarted","Data":"09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd"} Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.148689 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7v5fb" podStartSLOduration=3.016950772 podStartE2EDuration="46.148659711s" podCreationTimestamp="2026-02-20 06:43:09 +0000 UTC" firstStartedPulling="2026-02-20 06:43:11.458488483 +0000 UTC m=+148.229777462" lastFinishedPulling="2026-02-20 06:43:54.590197424 +0000 UTC m=+191.361486401" observedRunningTime="2026-02-20 06:43:55.14310269 +0000 UTC m=+191.914391668" watchObservedRunningTime="2026-02-20 06:43:55.148659711 +0000 UTC m=+191.919948689" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.215287 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 06:43:55 crc kubenswrapper[4492]: E0220 06:43:55.215598 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="extract-utilities" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.215617 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="extract-utilities" Feb 20 06:43:55 crc kubenswrapper[4492]: E0220 06:43:55.215628 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4ffa8d-dace-4d21-9774-1ca614e5671e" containerName="pruner" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.215635 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4ffa8d-dace-4d21-9774-1ca614e5671e" containerName="pruner" Feb 20 06:43:55 crc kubenswrapper[4492]: E0220 06:43:55.215642 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="registry-server" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.215648 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="registry-server" Feb 20 06:43:55 crc kubenswrapper[4492]: E0220 06:43:55.215664 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="extract-content" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.215671 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="extract-content" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.215781 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4ffa8d-dace-4d21-9774-1ca614e5671e" containerName="pruner" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.215795 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba7350b-1379-405e-918c-022f4b2fd1b5" containerName="registry-server" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.216273 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.219170 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.219306 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.228695 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.278733 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-var-lock\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.278929 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/deea33e9-6b00-41bd-af9a-4b4d900db00e-kube-api-access\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.279591 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.381254 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/deea33e9-6b00-41bd-af9a-4b4d900db00e-kube-api-access\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.381314 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.381378 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-var-lock\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.381395 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.381486 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-var-lock\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.397052 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/deea33e9-6b00-41bd-af9a-4b4d900db00e-kube-api-access\") pod \"installer-9-crc\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.570891 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:43:55 crc kubenswrapper[4492]: I0220 06:43:55.952640 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 20 06:43:55 crc kubenswrapper[4492]: W0220 06:43:55.959611 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poddeea33e9_6b00_41bd_af9a_4b4d900db00e.slice/crio-622c2ec4340bc398c79be911cfc3cbb210aa8b49a2934464537ccfb293f5fad7 WatchSource:0}: Error finding container 622c2ec4340bc398c79be911cfc3cbb210aa8b49a2934464537ccfb293f5fad7: Status 404 returned error can't find the container with id 622c2ec4340bc398c79be911cfc3cbb210aa8b49a2934464537ccfb293f5fad7 Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.141035 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"deea33e9-6b00-41bd-af9a-4b4d900db00e","Type":"ContainerStarted","Data":"622c2ec4340bc398c79be911cfc3cbb210aa8b49a2934464537ccfb293f5fad7"} Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.143321 4492 generic.go:334] "Generic (PLEG): container finished" podID="563da73d-0987-4fef-93f2-acf2e5755392" containerID="090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6" exitCode=0 Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.143381 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lpshv" event={"ID":"563da73d-0987-4fef-93f2-acf2e5755392","Type":"ContainerDied","Data":"090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6"} Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.147357 4492 generic.go:334] "Generic (PLEG): container finished" podID="23a7cdee-085f-4474-93f1-3edaea844222" containerID="6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a" exitCode=0 Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.147530 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prdt" event={"ID":"23a7cdee-085f-4474-93f1-3edaea844222","Type":"ContainerDied","Data":"6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a"} Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.889415 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.889778 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:56 crc kubenswrapper[4492]: I0220 06:43:56.933284 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:43:57 crc kubenswrapper[4492]: I0220 06:43:57.155850 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lpshv" event={"ID":"563da73d-0987-4fef-93f2-acf2e5755392","Type":"ContainerStarted","Data":"589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312"} Feb 20 06:43:57 crc kubenswrapper[4492]: I0220 06:43:57.157904 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prdt" event={"ID":"23a7cdee-085f-4474-93f1-3edaea844222","Type":"ContainerStarted","Data":"263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9"} Feb 20 06:43:57 crc kubenswrapper[4492]: I0220 06:43:57.159142 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"deea33e9-6b00-41bd-af9a-4b4d900db00e","Type":"ContainerStarted","Data":"902d916cec239c4307088cc36f6eb3db674addab73f912f83c2d7f4555dd14c2"} Feb 20 06:43:57 crc kubenswrapper[4492]: I0220 06:43:57.179343 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lpshv" podStartSLOduration=2.550618318 podStartE2EDuration="51.179321773s" podCreationTimestamp="2026-02-20 06:43:06 +0000 UTC" firstStartedPulling="2026-02-20 06:43:08.009709661 +0000 UTC m=+144.780998638" lastFinishedPulling="2026-02-20 06:43:56.638413115 +0000 UTC m=+193.409702093" observedRunningTime="2026-02-20 06:43:57.176670317 +0000 UTC m=+193.947959295" watchObservedRunningTime="2026-02-20 06:43:57.179321773 +0000 UTC m=+193.950610751" Feb 20 06:43:57 crc kubenswrapper[4492]: I0220 06:43:57.196613 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8prdt" podStartSLOduration=1.6274524970000002 podStartE2EDuration="49.196580507s" podCreationTimestamp="2026-02-20 06:43:08 +0000 UTC" firstStartedPulling="2026-02-20 06:43:09.148559149 +0000 UTC m=+145.919848126" lastFinishedPulling="2026-02-20 06:43:56.717687167 +0000 UTC m=+193.488976136" observedRunningTime="2026-02-20 06:43:57.193668509 +0000 UTC m=+193.964957486" watchObservedRunningTime="2026-02-20 06:43:57.196580507 +0000 UTC m=+193.967869485" Feb 20 06:43:58 crc kubenswrapper[4492]: I0220 06:43:58.441239 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:58 crc kubenswrapper[4492]: I0220 06:43:58.441316 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:58 crc kubenswrapper[4492]: I0220 06:43:58.475716 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:43:58 crc kubenswrapper[4492]: I0220 06:43:58.490205 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.490193314 podStartE2EDuration="3.490193314s" podCreationTimestamp="2026-02-20 06:43:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:43:57.212156192 +0000 UTC m=+193.983445170" watchObservedRunningTime="2026-02-20 06:43:58.490193314 +0000 UTC m=+195.261482291" Feb 20 06:43:58 crc kubenswrapper[4492]: I0220 06:43:58.857877 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:58 crc kubenswrapper[4492]: I0220 06:43:58.857942 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:58 crc kubenswrapper[4492]: I0220 06:43:58.911243 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:59 crc kubenswrapper[4492]: I0220 06:43:59.202428 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:43:59 crc kubenswrapper[4492]: I0220 06:43:59.483391 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:43:59 crc kubenswrapper[4492]: I0220 06:43:59.483939 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:44:00 crc kubenswrapper[4492]: I0220 06:44:00.518277 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7v5fb" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="registry-server" probeResult="failure" output=< Feb 20 06:44:00 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 06:44:00 crc kubenswrapper[4492]: > Feb 20 06:44:01 crc kubenswrapper[4492]: I0220 06:44:01.747493 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqsgt"] Feb 20 06:44:01 crc kubenswrapper[4492]: I0220 06:44:01.747721 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tqsgt" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="registry-server" containerID="cri-o://04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7" gracePeriod=2 Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.061196 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.162379 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-catalog-content\") pod \"25b6185c-71ad-476a-9dc9-97a5beeec215\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.162510 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kp6l7\" (UniqueName: \"kubernetes.io/projected/25b6185c-71ad-476a-9dc9-97a5beeec215-kube-api-access-kp6l7\") pod \"25b6185c-71ad-476a-9dc9-97a5beeec215\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.162606 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-utilities\") pod \"25b6185c-71ad-476a-9dc9-97a5beeec215\" (UID: \"25b6185c-71ad-476a-9dc9-97a5beeec215\") " Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.163712 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-utilities" (OuterVolumeSpecName: "utilities") pod "25b6185c-71ad-476a-9dc9-97a5beeec215" (UID: "25b6185c-71ad-476a-9dc9-97a5beeec215"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.168071 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b6185c-71ad-476a-9dc9-97a5beeec215-kube-api-access-kp6l7" (OuterVolumeSpecName: "kube-api-access-kp6l7") pod "25b6185c-71ad-476a-9dc9-97a5beeec215" (UID: "25b6185c-71ad-476a-9dc9-97a5beeec215"). InnerVolumeSpecName "kube-api-access-kp6l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.186788 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25b6185c-71ad-476a-9dc9-97a5beeec215" (UID: "25b6185c-71ad-476a-9dc9-97a5beeec215"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.191963 4492 generic.go:334] "Generic (PLEG): container finished" podID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerID="04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7" exitCode=0 Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.192018 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqsgt" event={"ID":"25b6185c-71ad-476a-9dc9-97a5beeec215","Type":"ContainerDied","Data":"04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7"} Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.192066 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tqsgt" event={"ID":"25b6185c-71ad-476a-9dc9-97a5beeec215","Type":"ContainerDied","Data":"6109633b47100ddc239ba70f201bdcbe643a60744b5c0f89b96e53a0d51f53a8"} Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.192088 4492 scope.go:117] "RemoveContainer" containerID="04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.192137 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tqsgt" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.220405 4492 scope.go:117] "RemoveContainer" containerID="b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.225034 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqsgt"] Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.227461 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tqsgt"] Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.235415 4492 scope.go:117] "RemoveContainer" containerID="a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.249544 4492 scope.go:117] "RemoveContainer" containerID="04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7" Feb 20 06:44:02 crc kubenswrapper[4492]: E0220 06:44:02.249897 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7\": container with ID starting with 04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7 not found: ID does not exist" containerID="04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.250141 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7"} err="failed to get container status \"04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7\": rpc error: code = NotFound desc = could not find container \"04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7\": container with ID starting with 04b346f7077c62d2f97c48a7576da2edfd13bd74b167cb95b77743a656358be7 not found: ID does not exist" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.250209 4492 scope.go:117] "RemoveContainer" containerID="b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d" Feb 20 06:44:02 crc kubenswrapper[4492]: E0220 06:44:02.250679 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d\": container with ID starting with b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d not found: ID does not exist" containerID="b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.250703 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d"} err="failed to get container status \"b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d\": rpc error: code = NotFound desc = could not find container \"b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d\": container with ID starting with b3072e7e2cde11b77adc55a7a882289732e276c10eca3974b3fde5e67d1ed80d not found: ID does not exist" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.250716 4492 scope.go:117] "RemoveContainer" containerID="a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419" Feb 20 06:44:02 crc kubenswrapper[4492]: E0220 06:44:02.251019 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419\": container with ID starting with a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419 not found: ID does not exist" containerID="a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.251058 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419"} err="failed to get container status \"a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419\": rpc error: code = NotFound desc = could not find container \"a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419\": container with ID starting with a581b75b9bb4273a52da2c7440fa4b75e72a3fcd9c2e950bfdceb97648adb419 not found: ID does not exist" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.264192 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.264216 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b6185c-71ad-476a-9dc9-97a5beeec215-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:02 crc kubenswrapper[4492]: I0220 06:44:02.264226 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kp6l7\" (UniqueName: \"kubernetes.io/projected/25b6185c-71ad-476a-9dc9-97a5beeec215-kube-api-access-kp6l7\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:03 crc kubenswrapper[4492]: I0220 06:44:03.564180 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" path="/var/lib/kubelet/pods/25b6185c-71ad-476a-9dc9-97a5beeec215/volumes" Feb 20 06:44:06 crc kubenswrapper[4492]: I0220 06:44:06.495933 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:44:06 crc kubenswrapper[4492]: I0220 06:44:06.496235 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:44:06 crc kubenswrapper[4492]: I0220 06:44:06.525935 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:44:06 crc kubenswrapper[4492]: I0220 06:44:06.918135 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.266537 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.568171 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f6t4g"] Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.568352 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f6t4g" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="registry-server" containerID="cri-o://812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7" gracePeriod=2 Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.868292 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.939485 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb48k\" (UniqueName: \"kubernetes.io/projected/7e865a3a-b443-4e73-a75c-24980c388b6f-kube-api-access-pb48k\") pod \"7e865a3a-b443-4e73-a75c-24980c388b6f\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.939622 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-utilities\") pod \"7e865a3a-b443-4e73-a75c-24980c388b6f\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.939651 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-catalog-content\") pod \"7e865a3a-b443-4e73-a75c-24980c388b6f\" (UID: \"7e865a3a-b443-4e73-a75c-24980c388b6f\") " Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.941983 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-utilities" (OuterVolumeSpecName: "utilities") pod "7e865a3a-b443-4e73-a75c-24980c388b6f" (UID: "7e865a3a-b443-4e73-a75c-24980c388b6f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.944325 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e865a3a-b443-4e73-a75c-24980c388b6f-kube-api-access-pb48k" (OuterVolumeSpecName: "kube-api-access-pb48k") pod "7e865a3a-b443-4e73-a75c-24980c388b6f" (UID: "7e865a3a-b443-4e73-a75c-24980c388b6f"). InnerVolumeSpecName "kube-api-access-pb48k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:07 crc kubenswrapper[4492]: I0220 06:44:07.983523 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e865a3a-b443-4e73-a75c-24980c388b6f" (UID: "7e865a3a-b443-4e73-a75c-24980c388b6f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.043563 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.043595 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e865a3a-b443-4e73-a75c-24980c388b6f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.043606 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb48k\" (UniqueName: \"kubernetes.io/projected/7e865a3a-b443-4e73-a75c-24980c388b6f-kube-api-access-pb48k\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.246181 4492 generic.go:334] "Generic (PLEG): container finished" podID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerID="812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7" exitCode=0 Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.246254 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t4g" event={"ID":"7e865a3a-b443-4e73-a75c-24980c388b6f","Type":"ContainerDied","Data":"812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7"} Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.246298 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f6t4g" event={"ID":"7e865a3a-b443-4e73-a75c-24980c388b6f","Type":"ContainerDied","Data":"3c0ce24fc48d9d7682539c147a7c1d4f3c6239664dce61365fb72b728ffb65a7"} Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.246320 4492 scope.go:117] "RemoveContainer" containerID="812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.246263 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f6t4g" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.263453 4492 scope.go:117] "RemoveContainer" containerID="555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.270511 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f6t4g"] Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.277073 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f6t4g"] Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.286833 4492 scope.go:117] "RemoveContainer" containerID="10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.299952 4492 scope.go:117] "RemoveContainer" containerID="812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7" Feb 20 06:44:08 crc kubenswrapper[4492]: E0220 06:44:08.300318 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7\": container with ID starting with 812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7 not found: ID does not exist" containerID="812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.300380 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7"} err="failed to get container status \"812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7\": rpc error: code = NotFound desc = could not find container \"812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7\": container with ID starting with 812b359fe527f5716a38f89fde9d4a7c6b2689474f28b0484c8bbb70008c88a7 not found: ID does not exist" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.300421 4492 scope.go:117] "RemoveContainer" containerID="555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930" Feb 20 06:44:08 crc kubenswrapper[4492]: E0220 06:44:08.300853 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930\": container with ID starting with 555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930 not found: ID does not exist" containerID="555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.300901 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930"} err="failed to get container status \"555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930\": rpc error: code = NotFound desc = could not find container \"555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930\": container with ID starting with 555aa6832c30a06dc45204220f69c9b7bf8fe66d2d2c014ac19f204da4642930 not found: ID does not exist" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.300929 4492 scope.go:117] "RemoveContainer" containerID="10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87" Feb 20 06:44:08 crc kubenswrapper[4492]: E0220 06:44:08.301187 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87\": container with ID starting with 10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87 not found: ID does not exist" containerID="10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.301213 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87"} err="failed to get container status \"10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87\": rpc error: code = NotFound desc = could not find container \"10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87\": container with ID starting with 10dfec2ce6eeb00062d17be8ced7a41753cdcd71a809d9d1ac3b5da8ccfb9e87 not found: ID does not exist" Feb 20 06:44:08 crc kubenswrapper[4492]: I0220 06:44:08.470112 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.311798 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.312248 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.312314 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.312932 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.313002 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf" gracePeriod=600 Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.517702 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.554076 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:44:09 crc kubenswrapper[4492]: I0220 06:44:09.564693 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" path="/var/lib/kubelet/pods/7e865a3a-b443-4e73-a75c-24980c388b6f/volumes" Feb 20 06:44:10 crc kubenswrapper[4492]: I0220 06:44:10.264253 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf" exitCode=0 Feb 20 06:44:10 crc kubenswrapper[4492]: I0220 06:44:10.264337 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf"} Feb 20 06:44:10 crc kubenswrapper[4492]: I0220 06:44:10.264898 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"8f0bbbbef7d028667e57fa52189868db9e398ca3ba68113428d3adebbf972641"} Feb 20 06:44:12 crc kubenswrapper[4492]: I0220 06:44:12.889724 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" podUID="baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" containerName="oauth-openshift" containerID="cri-o://15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd" gracePeriod=15 Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.176695 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.224756 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-router-certs\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.224800 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-login\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.224835 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-provider-selection\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.224858 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-trusted-ca-bundle\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.224908 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-serving-cert\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.224949 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-session\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225004 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-cliconfig\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225028 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-ocp-branding-template\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225045 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-idp-0-file-data\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225061 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-policies\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225086 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l6kp\" (UniqueName: \"kubernetes.io/projected/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-kube-api-access-6l6kp\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225100 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-error\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225123 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-dir\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225142 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-service-ca\") pod \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\" (UID: \"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3\") " Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.225947 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.226970 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.227309 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.228088 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.228108 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.231731 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-kube-api-access-6l6kp" (OuterVolumeSpecName: "kube-api-access-6l6kp") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "kube-api-access-6l6kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.231842 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.233817 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.234152 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.234458 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.235116 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.235210 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.238250 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.240821 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" (UID: "baa4cff7-d3ce-4abe-8d9b-56d1a945bce3"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.284077 4492 generic.go:334] "Generic (PLEG): container finished" podID="baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" containerID="15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd" exitCode=0 Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.284128 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" event={"ID":"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3","Type":"ContainerDied","Data":"15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd"} Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.284160 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" event={"ID":"baa4cff7-d3ce-4abe-8d9b-56d1a945bce3","Type":"ContainerDied","Data":"8e1419ba2eb50804e085056f5c1bc9c76a604561e03e02387bbfc98d42f6aebb"} Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.284165 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-p8dlg" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.284178 4492 scope.go:117] "RemoveContainer" containerID="15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.309829 4492 scope.go:117] "RemoveContainer" containerID="15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd" Feb 20 06:44:13 crc kubenswrapper[4492]: E0220 06:44:13.310950 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd\": container with ID starting with 15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd not found: ID does not exist" containerID="15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.310999 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd"} err="failed to get container status \"15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd\": rpc error: code = NotFound desc = could not find container \"15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd\": container with ID starting with 15eea3c19d8930ed6d107d4959972971643915127a174cbed3fb28ff293c21fd not found: ID does not exist" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.317715 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p8dlg"] Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.318257 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-p8dlg"] Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.326916 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.326946 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.326959 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.326972 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.326984 4492 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.326996 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l6kp\" (UniqueName: \"kubernetes.io/projected/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-kube-api-access-6l6kp\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327004 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327016 4492 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327026 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327035 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327043 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327053 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327063 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.327072 4492 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:13 crc kubenswrapper[4492]: I0220 06:44:13.563711 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" path="/var/lib/kubelet/pods/baa4cff7-d3ce-4abe-8d9b-56d1a945bce3/volumes" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.525635 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6775b6d8cc-4595h"] Feb 20 06:44:14 crc kubenswrapper[4492]: E0220 06:44:14.526213 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="extract-content" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526227 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="extract-content" Feb 20 06:44:14 crc kubenswrapper[4492]: E0220 06:44:14.526241 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" containerName="oauth-openshift" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526246 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" containerName="oauth-openshift" Feb 20 06:44:14 crc kubenswrapper[4492]: E0220 06:44:14.526258 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="registry-server" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526267 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="registry-server" Feb 20 06:44:14 crc kubenswrapper[4492]: E0220 06:44:14.526276 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="extract-utilities" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526282 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="extract-utilities" Feb 20 06:44:14 crc kubenswrapper[4492]: E0220 06:44:14.526290 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="extract-content" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526295 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="extract-content" Feb 20 06:44:14 crc kubenswrapper[4492]: E0220 06:44:14.526302 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="registry-server" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526308 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="registry-server" Feb 20 06:44:14 crc kubenswrapper[4492]: E0220 06:44:14.526315 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="extract-utilities" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526321 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="extract-utilities" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526427 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="baa4cff7-d3ce-4abe-8d9b-56d1a945bce3" containerName="oauth-openshift" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526438 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b6185c-71ad-476a-9dc9-97a5beeec215" containerName="registry-server" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526445 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e865a3a-b443-4e73-a75c-24980c388b6f" containerName="registry-server" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.526913 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.530004 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.530851 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.531001 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.531134 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.531374 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.531501 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.531951 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.532114 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.532555 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.532728 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.533186 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.542006 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.542410 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.547031 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.551139 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.560565 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6775b6d8cc-4595h"] Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.642840 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcbm2\" (UniqueName: \"kubernetes.io/projected/b754bbbc-5b66-40ad-89d8-31544cc94802-kube-api-access-lcbm2\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.642877 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-audit-policies\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.642904 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.642928 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-session\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.642948 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-service-ca\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.642967 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-router-certs\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.642996 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-error\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.643012 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.643033 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b754bbbc-5b66-40ad-89d8-31544cc94802-audit-dir\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.643056 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.643078 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.643095 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-login\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.643142 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.643164 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.744703 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.744866 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.744991 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-login\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.745131 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.745265 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749599 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcbm2\" (UniqueName: \"kubernetes.io/projected/b754bbbc-5b66-40ad-89d8-31544cc94802-kube-api-access-lcbm2\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749657 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-audit-policies\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749692 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749726 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-session\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749749 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-service-ca\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749771 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-router-certs\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749799 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-error\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749822 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749842 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b754bbbc-5b66-40ad-89d8-31544cc94802-audit-dir\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.749943 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b754bbbc-5b66-40ad-89d8-31544cc94802-audit-dir\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.754024 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.754059 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-audit-policies\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.754493 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-service-ca\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.755320 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-login\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.755508 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.758326 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.765835 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-template-error\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.766814 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.767553 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.767793 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-session\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.767928 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-router-certs\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.769285 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b754bbbc-5b66-40ad-89d8-31544cc94802-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.770319 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcbm2\" (UniqueName: \"kubernetes.io/projected/b754bbbc-5b66-40ad-89d8-31544cc94802-kube-api-access-lcbm2\") pod \"oauth-openshift-6775b6d8cc-4595h\" (UID: \"b754bbbc-5b66-40ad-89d8-31544cc94802\") " pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:14 crc kubenswrapper[4492]: I0220 06:44:14.845463 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:15 crc kubenswrapper[4492]: W0220 06:44:15.022845 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb754bbbc_5b66_40ad_89d8_31544cc94802.slice/crio-05261d313c3669bbcf8c7588cb99643a5da0482bced1dd94c19caadaf6b2207b WatchSource:0}: Error finding container 05261d313c3669bbcf8c7588cb99643a5da0482bced1dd94c19caadaf6b2207b: Status 404 returned error can't find the container with id 05261d313c3669bbcf8c7588cb99643a5da0482bced1dd94c19caadaf6b2207b Feb 20 06:44:15 crc kubenswrapper[4492]: I0220 06:44:15.034164 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6775b6d8cc-4595h"] Feb 20 06:44:15 crc kubenswrapper[4492]: I0220 06:44:15.297461 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" event={"ID":"b754bbbc-5b66-40ad-89d8-31544cc94802","Type":"ContainerStarted","Data":"20ce461317e9aa0495189ecf61867e2724ea3a6d0edc19e4a89c10ba762b9076"} Feb 20 06:44:15 crc kubenswrapper[4492]: I0220 06:44:15.297529 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" event={"ID":"b754bbbc-5b66-40ad-89d8-31544cc94802","Type":"ContainerStarted","Data":"05261d313c3669bbcf8c7588cb99643a5da0482bced1dd94c19caadaf6b2207b"} Feb 20 06:44:15 crc kubenswrapper[4492]: I0220 06:44:15.298689 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:15 crc kubenswrapper[4492]: I0220 06:44:15.316747 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" podStartSLOduration=28.316726434 podStartE2EDuration="28.316726434s" podCreationTimestamp="2026-02-20 06:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:44:15.313066172 +0000 UTC m=+212.084355160" watchObservedRunningTime="2026-02-20 06:44:15.316726434 +0000 UTC m=+212.088015412" Feb 20 06:44:15 crc kubenswrapper[4492]: I0220 06:44:15.543872 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6775b6d8cc-4595h" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.395878 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lpshv"] Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.398851 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lpshv" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="registry-server" containerID="cri-o://589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312" gracePeriod=30 Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.400017 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9f62m"] Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.400401 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9f62m" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="registry-server" containerID="cri-o://151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271" gracePeriod=30 Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.417761 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5snjq"] Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.418016 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" podUID="70f428c8-e365-42aa-bced-3c9b721b6c90" containerName="marketplace-operator" containerID="cri-o://13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305" gracePeriod=30 Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.421617 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prdt"] Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.421863 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8prdt" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="registry-server" containerID="cri-o://263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9" gracePeriod=30 Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.433281 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7v5fb"] Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.433488 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7v5fb" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="registry-server" containerID="cri-o://09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd" gracePeriod=30 Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.435289 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddlzq"] Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.437667 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.447605 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddlzq"] Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.461764 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99c4s\" (UniqueName: \"kubernetes.io/projected/1be6eb66-d6a5-41a8-acf9-8297fefe0279-kube-api-access-99c4s\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.462053 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1be6eb66-d6a5-41a8-acf9-8297fefe0279-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.462076 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1be6eb66-d6a5-41a8-acf9-8297fefe0279-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.562626 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1be6eb66-d6a5-41a8-acf9-8297fefe0279-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.562665 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1be6eb66-d6a5-41a8-acf9-8297fefe0279-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.562712 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99c4s\" (UniqueName: \"kubernetes.io/projected/1be6eb66-d6a5-41a8-acf9-8297fefe0279-kube-api-access-99c4s\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.567238 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1be6eb66-d6a5-41a8-acf9-8297fefe0279-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.583086 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/1be6eb66-d6a5-41a8-acf9-8297fefe0279-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.584620 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99c4s\" (UniqueName: \"kubernetes.io/projected/1be6eb66-d6a5-41a8-acf9-8297fefe0279-kube-api-access-99c4s\") pod \"marketplace-operator-79b997595-ddlzq\" (UID: \"1be6eb66-d6a5-41a8-acf9-8297fefe0279\") " pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.744467 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.778810 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.861632 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.867226 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-utilities\") pod \"563da73d-0987-4fef-93f2-acf2e5755392\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.867309 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klrjk\" (UniqueName: \"kubernetes.io/projected/563da73d-0987-4fef-93f2-acf2e5755392-kube-api-access-klrjk\") pod \"563da73d-0987-4fef-93f2-acf2e5755392\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.867368 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-catalog-content\") pod \"563da73d-0987-4fef-93f2-acf2e5755392\" (UID: \"563da73d-0987-4fef-93f2-acf2e5755392\") " Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.868298 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-utilities" (OuterVolumeSpecName: "utilities") pod "563da73d-0987-4fef-93f2-acf2e5755392" (UID: "563da73d-0987-4fef-93f2-acf2e5755392"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.870606 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563da73d-0987-4fef-93f2-acf2e5755392-kube-api-access-klrjk" (OuterVolumeSpecName: "kube-api-access-klrjk") pod "563da73d-0987-4fef-93f2-acf2e5755392" (UID: "563da73d-0987-4fef-93f2-acf2e5755392"). InnerVolumeSpecName "kube-api-access-klrjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.904752 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.926613 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "563da73d-0987-4fef-93f2-acf2e5755392" (UID: "563da73d-0987-4fef-93f2-acf2e5755392"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.938409 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.939018 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.971667 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nfdb\" (UniqueName: \"kubernetes.io/projected/8c489688-3ccb-4ad8-b309-da0d10f50618-kube-api-access-7nfdb\") pod \"8c489688-3ccb-4ad8-b309-da0d10f50618\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.971768 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-catalog-content\") pod \"8c489688-3ccb-4ad8-b309-da0d10f50618\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.971887 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-utilities\") pod \"8c489688-3ccb-4ad8-b309-da0d10f50618\" (UID: \"8c489688-3ccb-4ad8-b309-da0d10f50618\") " Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.972202 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.972216 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klrjk\" (UniqueName: \"kubernetes.io/projected/563da73d-0987-4fef-93f2-acf2e5755392-kube-api-access-klrjk\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.972226 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/563da73d-0987-4fef-93f2-acf2e5755392-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.972822 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-utilities" (OuterVolumeSpecName: "utilities") pod "8c489688-3ccb-4ad8-b309-da0d10f50618" (UID: "8c489688-3ccb-4ad8-b309-da0d10f50618"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:31 crc kubenswrapper[4492]: I0220 06:44:31.978445 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c489688-3ccb-4ad8-b309-da0d10f50618-kube-api-access-7nfdb" (OuterVolumeSpecName: "kube-api-access-7nfdb") pod "8c489688-3ccb-4ad8-b309-da0d10f50618" (UID: "8c489688-3ccb-4ad8-b309-da0d10f50618"). InnerVolumeSpecName "kube-api-access-7nfdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.048970 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c489688-3ccb-4ad8-b309-da0d10f50618" (UID: "8c489688-3ccb-4ad8-b309-da0d10f50618"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.072656 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7xsj\" (UniqueName: \"kubernetes.io/projected/23a7cdee-085f-4474-93f1-3edaea844222-kube-api-access-z7xsj\") pod \"23a7cdee-085f-4474-93f1-3edaea844222\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.072741 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pzxg\" (UniqueName: \"kubernetes.io/projected/70f428c8-e365-42aa-bced-3c9b721b6c90-kube-api-access-6pzxg\") pod \"70f428c8-e365-42aa-bced-3c9b721b6c90\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.072772 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-trusted-ca\") pod \"70f428c8-e365-42aa-bced-3c9b721b6c90\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.072812 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-operator-metrics\") pod \"70f428c8-e365-42aa-bced-3c9b721b6c90\" (UID: \"70f428c8-e365-42aa-bced-3c9b721b6c90\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.072891 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-catalog-content\") pod \"23a7cdee-085f-4474-93f1-3edaea844222\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.072921 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-catalog-content\") pod \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.072972 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-utilities\") pod \"23a7cdee-085f-4474-93f1-3edaea844222\" (UID: \"23a7cdee-085f-4474-93f1-3edaea844222\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.073010 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-utilities\") pod \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.073038 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r9hn\" (UniqueName: \"kubernetes.io/projected/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-kube-api-access-6r9hn\") pod \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\" (UID: \"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139\") " Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.073394 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nfdb\" (UniqueName: \"kubernetes.io/projected/8c489688-3ccb-4ad8-b309-da0d10f50618-kube-api-access-7nfdb\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.073407 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.073418 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c489688-3ccb-4ad8-b309-da0d10f50618-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.074180 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-utilities" (OuterVolumeSpecName: "utilities") pod "2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" (UID: "2fcf3fe3-c5f4-40a4-8b92-06d408c9d139"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.075309 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-utilities" (OuterVolumeSpecName: "utilities") pod "23a7cdee-085f-4474-93f1-3edaea844222" (UID: "23a7cdee-085f-4474-93f1-3edaea844222"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.076457 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "70f428c8-e365-42aa-bced-3c9b721b6c90" (UID: "70f428c8-e365-42aa-bced-3c9b721b6c90"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.080644 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23a7cdee-085f-4474-93f1-3edaea844222-kube-api-access-z7xsj" (OuterVolumeSpecName: "kube-api-access-z7xsj") pod "23a7cdee-085f-4474-93f1-3edaea844222" (UID: "23a7cdee-085f-4474-93f1-3edaea844222"). InnerVolumeSpecName "kube-api-access-z7xsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.096366 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-kube-api-access-6r9hn" (OuterVolumeSpecName: "kube-api-access-6r9hn") pod "2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" (UID: "2fcf3fe3-c5f4-40a4-8b92-06d408c9d139"). InnerVolumeSpecName "kube-api-access-6r9hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.096801 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f428c8-e365-42aa-bced-3c9b721b6c90-kube-api-access-6pzxg" (OuterVolumeSpecName: "kube-api-access-6pzxg") pod "70f428c8-e365-42aa-bced-3c9b721b6c90" (UID: "70f428c8-e365-42aa-bced-3c9b721b6c90"). InnerVolumeSpecName "kube-api-access-6pzxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.105891 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23a7cdee-085f-4474-93f1-3edaea844222" (UID: "23a7cdee-085f-4474-93f1-3edaea844222"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.107269 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "70f428c8-e365-42aa-bced-3c9b721b6c90" (UID: "70f428c8-e365-42aa-bced-3c9b721b6c90"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.135810 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ddlzq"] Feb 20 06:44:32 crc kubenswrapper[4492]: W0220 06:44:32.143517 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1be6eb66_d6a5_41a8_acf9_8297fefe0279.slice/crio-b2378bcd0dcb8611a7e208a30d074e7d1983f3d893e5a1e663e6fb7e574e6452 WatchSource:0}: Error finding container b2378bcd0dcb8611a7e208a30d074e7d1983f3d893e5a1e663e6fb7e574e6452: Status 404 returned error can't find the container with id b2378bcd0dcb8611a7e208a30d074e7d1983f3d893e5a1e663e6fb7e574e6452 Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177522 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pzxg\" (UniqueName: \"kubernetes.io/projected/70f428c8-e365-42aa-bced-3c9b721b6c90-kube-api-access-6pzxg\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177567 4492 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177582 4492 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/70f428c8-e365-42aa-bced-3c9b721b6c90-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177596 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177607 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23a7cdee-085f-4474-93f1-3edaea844222-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177618 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177628 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r9hn\" (UniqueName: \"kubernetes.io/projected/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-kube-api-access-6r9hn\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.177638 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7xsj\" (UniqueName: \"kubernetes.io/projected/23a7cdee-085f-4474-93f1-3edaea844222-kube-api-access-z7xsj\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.221291 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" (UID: "2fcf3fe3-c5f4-40a4-8b92-06d408c9d139"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.278996 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.405329 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" event={"ID":"1be6eb66-d6a5-41a8-acf9-8297fefe0279","Type":"ContainerStarted","Data":"129340611aea01918bd10a53232cdc56c90d1c4fc7a48952edb502d8b56fe1b6"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.405384 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" event={"ID":"1be6eb66-d6a5-41a8-acf9-8297fefe0279","Type":"ContainerStarted","Data":"b2378bcd0dcb8611a7e208a30d074e7d1983f3d893e5a1e663e6fb7e574e6452"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.405645 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.407025 4492 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ddlzq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" start-of-body= Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.407079 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" podUID="1be6eb66-d6a5-41a8-acf9-8297fefe0279" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.409921 4492 generic.go:334] "Generic (PLEG): container finished" podID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerID="151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271" exitCode=0 Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.409986 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9f62m" event={"ID":"8c489688-3ccb-4ad8-b309-da0d10f50618","Type":"ContainerDied","Data":"151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.410013 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9f62m" event={"ID":"8c489688-3ccb-4ad8-b309-da0d10f50618","Type":"ContainerDied","Data":"6b47cb8104fda9aaf7fcad56b2e7c1caa7b56ad2d281bf40a38835eb50f4daa3"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.410039 4492 scope.go:117] "RemoveContainer" containerID="151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.410195 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9f62m" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.416175 4492 generic.go:334] "Generic (PLEG): container finished" podID="563da73d-0987-4fef-93f2-acf2e5755392" containerID="589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312" exitCode=0 Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.416257 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lpshv" event={"ID":"563da73d-0987-4fef-93f2-acf2e5755392","Type":"ContainerDied","Data":"589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.416285 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lpshv" event={"ID":"563da73d-0987-4fef-93f2-acf2e5755392","Type":"ContainerDied","Data":"4c144757d0342f75c034ad245564ad01479053efee212ac8503168b9afbc3e51"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.416375 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lpshv" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.422316 4492 generic.go:334] "Generic (PLEG): container finished" podID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerID="09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd" exitCode=0 Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.422414 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7v5fb" event={"ID":"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139","Type":"ContainerDied","Data":"09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.422451 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7v5fb" event={"ID":"2fcf3fe3-c5f4-40a4-8b92-06d408c9d139","Type":"ContainerDied","Data":"dfd4c907ad8803890d0e1df875eef3611251b457ef7ab7c8d5735395535def50"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.428399 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" podStartSLOduration=1.428374192 podStartE2EDuration="1.428374192s" podCreationTimestamp="2026-02-20 06:44:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:44:32.421395262 +0000 UTC m=+229.192684239" watchObservedRunningTime="2026-02-20 06:44:32.428374192 +0000 UTC m=+229.199663170" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.428724 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7v5fb" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.429384 4492 generic.go:334] "Generic (PLEG): container finished" podID="70f428c8-e365-42aa-bced-3c9b721b6c90" containerID="13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305" exitCode=0 Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.429455 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" event={"ID":"70f428c8-e365-42aa-bced-3c9b721b6c90","Type":"ContainerDied","Data":"13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.429553 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" event={"ID":"70f428c8-e365-42aa-bced-3c9b721b6c90","Type":"ContainerDied","Data":"4c9c16a85a042558187c01a69930886ac41c156ea570331bdb97de52df5982c2"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.429602 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5snjq" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.435074 4492 generic.go:334] "Generic (PLEG): container finished" podID="23a7cdee-085f-4474-93f1-3edaea844222" containerID="263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9" exitCode=0 Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.435210 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prdt" event={"ID":"23a7cdee-085f-4474-93f1-3edaea844222","Type":"ContainerDied","Data":"263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.435325 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prdt" event={"ID":"23a7cdee-085f-4474-93f1-3edaea844222","Type":"ContainerDied","Data":"b80ad750ef6b914fbe1049d856e3e7f536c134d4b07057a50b63755544ce4a94"} Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.436466 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prdt" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.446762 4492 scope.go:117] "RemoveContainer" containerID="96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.454019 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9f62m"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.458418 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9f62m"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.473852 4492 scope.go:117] "RemoveContainer" containerID="c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.476490 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lpshv"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.482241 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lpshv"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.486432 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7v5fb"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.498542 4492 scope.go:117] "RemoveContainer" containerID="151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.499410 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271\": container with ID starting with 151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271 not found: ID does not exist" containerID="151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.499497 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271"} err="failed to get container status \"151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271\": rpc error: code = NotFound desc = could not find container \"151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271\": container with ID starting with 151771ce3df77632072868f98ed1eba225303c1127b26badb3305e43f4669271 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.499581 4492 scope.go:117] "RemoveContainer" containerID="96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.500224 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2\": container with ID starting with 96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2 not found: ID does not exist" containerID="96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.501001 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2"} err="failed to get container status \"96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2\": rpc error: code = NotFound desc = could not find container \"96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2\": container with ID starting with 96130d678e9cfc373688cc6a6919f9931b5b1901b1cd392378a8603e04a1d1e2 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.501208 4492 scope.go:117] "RemoveContainer" containerID="c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.501591 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5\": container with ID starting with c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5 not found: ID does not exist" containerID="c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.501997 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5"} err="failed to get container status \"c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5\": rpc error: code = NotFound desc = could not find container \"c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5\": container with ID starting with c3d5a511be29eeefd3b774ceb06104b04678dcfc5bec300ee5d115e403c8aeb5 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.502268 4492 scope.go:117] "RemoveContainer" containerID="589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.524802 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7v5fb"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.528628 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prdt"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.530714 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prdt"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.533380 4492 scope.go:117] "RemoveContainer" containerID="090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.533493 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5snjq"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.534956 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5snjq"] Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.564852 4492 scope.go:117] "RemoveContainer" containerID="39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.579310 4492 scope.go:117] "RemoveContainer" containerID="589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.580287 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312\": container with ID starting with 589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312 not found: ID does not exist" containerID="589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.580328 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312"} err="failed to get container status \"589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312\": rpc error: code = NotFound desc = could not find container \"589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312\": container with ID starting with 589cfee79aa5a06b0dbb350cc524bf4fd9920c46bcbbd093b689b6a9c28ab312 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.580364 4492 scope.go:117] "RemoveContainer" containerID="090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.580770 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6\": container with ID starting with 090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6 not found: ID does not exist" containerID="090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.580916 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6"} err="failed to get container status \"090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6\": rpc error: code = NotFound desc = could not find container \"090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6\": container with ID starting with 090847f54cfce77fabf923904d28282e86166d68c0ab9ece3680705473752ef6 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.580941 4492 scope.go:117] "RemoveContainer" containerID="39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.582080 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c\": container with ID starting with 39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c not found: ID does not exist" containerID="39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.582114 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c"} err="failed to get container status \"39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c\": rpc error: code = NotFound desc = could not find container \"39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c\": container with ID starting with 39a1c44aff5f2b6e1a1281d43ffc26f5cc0dcd4bb4da05c48a9329a97a54929c not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.582136 4492 scope.go:117] "RemoveContainer" containerID="09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.596884 4492 scope.go:117] "RemoveContainer" containerID="7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.616833 4492 scope.go:117] "RemoveContainer" containerID="eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.631262 4492 scope.go:117] "RemoveContainer" containerID="09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.631784 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd\": container with ID starting with 09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd not found: ID does not exist" containerID="09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.631821 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd"} err="failed to get container status \"09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd\": rpc error: code = NotFound desc = could not find container \"09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd\": container with ID starting with 09da9cc80b22459619350dfed32aa4df0f56c384a26e23cb65c8c085c36701fd not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.631849 4492 scope.go:117] "RemoveContainer" containerID="7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.632365 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6\": container with ID starting with 7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6 not found: ID does not exist" containerID="7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.632432 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6"} err="failed to get container status \"7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6\": rpc error: code = NotFound desc = could not find container \"7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6\": container with ID starting with 7c469e2bda65758dd59105d8c6e0b8a8f2f60af5b10f62aefbb2bbb70ee848a6 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.632495 4492 scope.go:117] "RemoveContainer" containerID="eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.633904 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b\": container with ID starting with eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b not found: ID does not exist" containerID="eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.633948 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b"} err="failed to get container status \"eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b\": rpc error: code = NotFound desc = could not find container \"eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b\": container with ID starting with eefc8eb4faf13276955bf8e0ae231fffed12342fb382e9e4dad96a57ca3dc94b not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.633976 4492 scope.go:117] "RemoveContainer" containerID="13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.645996 4492 scope.go:117] "RemoveContainer" containerID="13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.646492 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305\": container with ID starting with 13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305 not found: ID does not exist" containerID="13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.646531 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305"} err="failed to get container status \"13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305\": rpc error: code = NotFound desc = could not find container \"13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305\": container with ID starting with 13022c614267d5f7d5f1934980b16b91ed43a55a0c2aef6333d3bda236f1f305 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.646549 4492 scope.go:117] "RemoveContainer" containerID="263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.663953 4492 scope.go:117] "RemoveContainer" containerID="6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.678286 4492 scope.go:117] "RemoveContainer" containerID="39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.712418 4492 scope.go:117] "RemoveContainer" containerID="263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.712791 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9\": container with ID starting with 263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9 not found: ID does not exist" containerID="263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.712818 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9"} err="failed to get container status \"263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9\": rpc error: code = NotFound desc = could not find container \"263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9\": container with ID starting with 263cc116a8af1e41a97ab95461f2e258f4afc9e2f60324e4827b5a0c7a6702a9 not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.712840 4492 scope.go:117] "RemoveContainer" containerID="6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.713192 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a\": container with ID starting with 6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a not found: ID does not exist" containerID="6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.713214 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a"} err="failed to get container status \"6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a\": rpc error: code = NotFound desc = could not find container \"6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a\": container with ID starting with 6236d94df92851cde3c3e99ef7315a8d99ebee9b65b362d8bcdd2bb23b4fc17a not found: ID does not exist" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.713231 4492 scope.go:117] "RemoveContainer" containerID="39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1" Feb 20 06:44:32 crc kubenswrapper[4492]: E0220 06:44:32.714742 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1\": container with ID starting with 39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1 not found: ID does not exist" containerID="39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1" Feb 20 06:44:32 crc kubenswrapper[4492]: I0220 06:44:32.714763 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1"} err="failed to get container status \"39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1\": rpc error: code = NotFound desc = could not find container \"39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1\": container with ID starting with 39f8e3498bef57318737c42ecd9002e80e6af0e9f202cb6b03e1946d8bbdbec1 not found: ID does not exist" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.462166 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ddlzq" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.564344 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23a7cdee-085f-4474-93f1-3edaea844222" path="/var/lib/kubelet/pods/23a7cdee-085f-4474-93f1-3edaea844222/volumes" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.565445 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" path="/var/lib/kubelet/pods/2fcf3fe3-c5f4-40a4-8b92-06d408c9d139/volumes" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.566186 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="563da73d-0987-4fef-93f2-acf2e5755392" path="/var/lib/kubelet/pods/563da73d-0987-4fef-93f2-acf2e5755392/volumes" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.568101 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f428c8-e365-42aa-bced-3c9b721b6c90" path="/var/lib/kubelet/pods/70f428c8-e365-42aa-bced-3c9b721b6c90/volumes" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.568929 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" path="/var/lib/kubelet/pods/8c489688-3ccb-4ad8-b309-da0d10f50618/volumes" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.613992 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cms44"] Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614347 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614374 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614389 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614395 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614408 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614416 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614424 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614431 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614442 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614450 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614461 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614467 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614498 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614505 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614513 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614521 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614535 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614543 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="extract-content" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614554 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614562 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614572 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614578 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614588 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f428c8-e365-42aa-bced-3c9b721b6c90" containerName="marketplace-operator" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614593 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f428c8-e365-42aa-bced-3c9b721b6c90" containerName="marketplace-operator" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.614601 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614609 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="extract-utilities" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614734 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f428c8-e365-42aa-bced-3c9b721b6c90" containerName="marketplace-operator" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614746 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fcf3fe3-c5f4-40a4-8b92-06d408c9d139" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614756 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c489688-3ccb-4ad8-b309-da0d10f50618" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614764 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="23a7cdee-085f-4474-93f1-3edaea844222" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.614772 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="563da73d-0987-4fef-93f2-acf2e5755392" containerName="registry-server" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.615827 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.617578 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.617781 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cms44"] Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.694936 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcd6b9a-5837-48eb-98d0-691541bfff4d-utilities\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.695038 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgzgr\" (UniqueName: \"kubernetes.io/projected/2bcd6b9a-5837-48eb-98d0-691541bfff4d-kube-api-access-dgzgr\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.695113 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcd6b9a-5837-48eb-98d0-691541bfff4d-catalog-content\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.796130 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgzgr\" (UniqueName: \"kubernetes.io/projected/2bcd6b9a-5837-48eb-98d0-691541bfff4d-kube-api-access-dgzgr\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.796354 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcd6b9a-5837-48eb-98d0-691541bfff4d-catalog-content\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.796562 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcd6b9a-5837-48eb-98d0-691541bfff4d-utilities\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.797274 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcd6b9a-5837-48eb-98d0-691541bfff4d-utilities\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.798823 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcd6b9a-5837-48eb-98d0-691541bfff4d-catalog-content\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.805888 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5kkhh"] Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.807305 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.809743 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.829884 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5kkhh"] Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.838234 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgzgr\" (UniqueName: \"kubernetes.io/projected/2bcd6b9a-5837-48eb-98d0-691541bfff4d-kube-api-access-dgzgr\") pod \"redhat-marketplace-cms44\" (UID: \"2bcd6b9a-5837-48eb-98d0-691541bfff4d\") " pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.898454 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.898960 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d2d17f0-7864-41dd-8b9f-285944a57376-catalog-content\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.899129 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d2d17f0-7864-41dd-8b9f-285944a57376-utilities\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.931031 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.936321 4492 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.937291 4492 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.937446 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.937736 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321" gracePeriod=15 Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.937757 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac" gracePeriod=15 Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.937790 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b" gracePeriod=15 Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.937846 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c" gracePeriod=15 Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.937848 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c" gracePeriod=15 Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938567 4492 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.938782 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938805 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.938815 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938821 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.938830 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938836 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.938844 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938850 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.938859 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938867 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.938874 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938880 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 20 06:44:33 crc kubenswrapper[4492]: E0220 06:44:33.938888 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938894 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938985 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.938995 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.939006 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.939014 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.939021 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 20 06:44:33 crc kubenswrapper[4492]: I0220 06:44:33.939028 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.002063 4492 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.26.25:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.005234 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d2d17f0-7864-41dd-8b9f-285944a57376-catalog-content\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.005279 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d2d17f0-7864-41dd-8b9f-285944a57376-utilities\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.005387 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.005785 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d2d17f0-7864-41dd-8b9f-285944a57376-catalog-content\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.005791 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d2d17f0-7864-41dd-8b9f-285944a57376-utilities\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.007727 4492 projected.go:194] Error preparing data for projected volume kube-api-access-59v8d for pod openshift-marketplace/redhat-operators-5kkhh: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.007801 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d podName:9d2d17f0-7864-41dd-8b9f-285944a57376 nodeName:}" failed. No retries permitted until 2026-02-20 06:44:34.50777815 +0000 UTC m=+231.279067119 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-59v8d" (UniqueName: "kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d") pod "redhat-operators-5kkhh" (UID: "9d2d17f0-7864-41dd-8b9f-285944a57376") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.008107 4492 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 192.168.26.25:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-5kkhh.1895e16a0b0441a0 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-5kkhh,UID:9d2d17f0-7864-41dd-8b9f-285944a57376,APIVersion:v1,ResourceVersion:29621,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-59v8d\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token\": dial tcp 192.168.26.25:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 06:44:34.007769504 +0000 UTC m=+230.779058482,LastTimestamp:2026-02-20 06:44:34.007769504 +0000 UTC m=+230.779058482,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.107139 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.107190 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.108645 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.108751 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.108906 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.109002 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.109055 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.109077 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.210865 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.210928 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.210966 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211013 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211093 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211152 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211184 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211212 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211321 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211376 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211406 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211430 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211457 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211496 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211530 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.211554 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.302570 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:34 crc kubenswrapper[4492]: W0220 06:44:34.320209 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-f8391974084fd69804432f74123049566feeda460c0b9dbbbaf7dc50a6f693ed WatchSource:0}: Error finding container f8391974084fd69804432f74123049566feeda460c0b9dbbbaf7dc50a6f693ed: Status 404 returned error can't find the container with id f8391974084fd69804432f74123049566feeda460c0b9dbbbaf7dc50a6f693ed Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.329582 4492 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 20 06:44:34 crc kubenswrapper[4492]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4" Netns:"/var/run/netns/c0aa2b68-e5b9-4d33-8a83-7ba97eb48113" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d" Path:"" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 20 06:44:34 crc kubenswrapper[4492]: > Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.329643 4492 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 20 06:44:34 crc kubenswrapper[4492]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4" Netns:"/var/run/netns/c0aa2b68-e5b9-4d33-8a83-7ba97eb48113" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d" Path:"" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 20 06:44:34 crc kubenswrapper[4492]: > pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.329666 4492 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Feb 20 06:44:34 crc kubenswrapper[4492]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4" Netns:"/var/run/netns/c0aa2b68-e5b9-4d33-8a83-7ba97eb48113" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d" Path:"" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 20 06:44:34 crc kubenswrapper[4492]: > pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.329726 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-cms44_openshift-marketplace(2bcd6b9a-5837-48eb-98d0-691541bfff4d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-cms44_openshift-marketplace(2bcd6b9a-5837-48eb-98d0-691541bfff4d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4\\\" Netns:\\\"/var/run/netns/c0aa2b68-e5b9-4d33-8a83-7ba97eb48113\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=b87365553e3b3b859d109d6e9a15182d99a041c3df2e0d47451a7842f62c00f4;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s\\\": dial tcp 192.168.26.25:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/redhat-marketplace-cms44" podUID="2bcd6b9a-5837-48eb-98d0-691541bfff4d" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.461147 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.463839 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.464621 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c" exitCode=0 Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.464656 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac" exitCode=0 Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.464668 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b" exitCode=0 Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.464675 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c" exitCode=2 Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.464750 4492 scope.go:117] "RemoveContainer" containerID="cc152aee2e829c277e027af110e23fe676a5fde6821b7d569fab0708c5c18e26" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.466870 4492 generic.go:334] "Generic (PLEG): container finished" podID="deea33e9-6b00-41bd-af9a-4b4d900db00e" containerID="902d916cec239c4307088cc36f6eb3db674addab73f912f83c2d7f4555dd14c2" exitCode=0 Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.466955 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"deea33e9-6b00-41bd-af9a-4b4d900db00e","Type":"ContainerDied","Data":"902d916cec239c4307088cc36f6eb3db674addab73f912f83c2d7f4555dd14c2"} Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.467897 4492 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.468400 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.468587 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"f8391974084fd69804432f74123049566feeda460c0b9dbbbaf7dc50a6f693ed"} Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.468811 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.469297 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:34 crc kubenswrapper[4492]: I0220 06:44:34.516860 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.517768 4492 projected.go:194] Error preparing data for projected volume kube-api-access-59v8d for pod openshift-marketplace/redhat-operators-5kkhh: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.517873 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d podName:9d2d17f0-7864-41dd-8b9f-285944a57376 nodeName:}" failed. No retries permitted until 2026-02-20 06:44:35.517833098 +0000 UTC m=+232.289122076 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-59v8d" (UniqueName: "kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d") pod "redhat-operators-5kkhh" (UID: "9d2d17f0-7864-41dd-8b9f-285944a57376") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.800003 4492 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 20 06:44:34 crc kubenswrapper[4492]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d" Netns:"/var/run/netns/1a174084-3208-4b18-9a79-06162bcf4da5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d" Path:"" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 20 06:44:34 crc kubenswrapper[4492]: > Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.800351 4492 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 20 06:44:34 crc kubenswrapper[4492]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d" Netns:"/var/run/netns/1a174084-3208-4b18-9a79-06162bcf4da5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d" Path:"" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 20 06:44:34 crc kubenswrapper[4492]: > pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.800390 4492 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Feb 20 06:44:34 crc kubenswrapper[4492]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d" Netns:"/var/run/netns/1a174084-3208-4b18-9a79-06162bcf4da5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d" Path:"" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:34 crc kubenswrapper[4492]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 20 06:44:34 crc kubenswrapper[4492]: > pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:34 crc kubenswrapper[4492]: E0220 06:44:34.800569 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redhat-marketplace-cms44_openshift-marketplace(2bcd6b9a-5837-48eb-98d0-691541bfff4d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redhat-marketplace-cms44_openshift-marketplace(2bcd6b9a-5837-48eb-98d0-691541bfff4d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redhat-marketplace-cms44_openshift-marketplace_2bcd6b9a-5837-48eb-98d0-691541bfff4d_0(eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d): error adding pod openshift-marketplace_redhat-marketplace-cms44 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d\\\" Netns:\\\"/var/run/netns/1a174084-3208-4b18-9a79-06162bcf4da5\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=redhat-marketplace-cms44;K8S_POD_INFRA_CONTAINER_ID=eab6445db80038e6a83135b4f9817e7b98eb15492f02f3c2d51badceb46e7a8d;K8S_POD_UID=2bcd6b9a-5837-48eb-98d0-691541bfff4d\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/redhat-marketplace-cms44] networking: Multus: [openshift-marketplace/redhat-marketplace-cms44/2bcd6b9a-5837-48eb-98d0-691541bfff4d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: SetNetworkStatus: failed to update the pod redhat-marketplace-cms44 in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-cms44?timeout=1m0s\\\": dial tcp 192.168.26.25:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/redhat-marketplace-cms44" podUID="2bcd6b9a-5837-48eb-98d0-691541bfff4d" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.481443 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"73c73a886d5369936df5a17ed652bcecc0dc75b8c03c83adaa1b6082e8f07c82"} Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.482666 4492 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:35 crc kubenswrapper[4492]: E0220 06:44:35.482706 4492 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.26.25:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.483139 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.485781 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.526074 4492 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.526133 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.542216 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:35 crc kubenswrapper[4492]: E0220 06:44:35.543176 4492 projected.go:194] Error preparing data for projected volume kube-api-access-59v8d for pod openshift-marketplace/redhat-operators-5kkhh: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:35 crc kubenswrapper[4492]: E0220 06:44:35.543294 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d podName:9d2d17f0-7864-41dd-8b9f-285944a57376 nodeName:}" failed. No retries permitted until 2026-02-20 06:44:37.543267729 +0000 UTC m=+234.314556707 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-59v8d" (UniqueName: "kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d") pod "redhat-operators-5kkhh" (UID: "9d2d17f0-7864-41dd-8b9f-285944a57376") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.685676 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.686454 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.846903 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/deea33e9-6b00-41bd-af9a-4b4d900db00e-kube-api-access\") pod \"deea33e9-6b00-41bd-af9a-4b4d900db00e\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.847158 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-kubelet-dir\") pod \"deea33e9-6b00-41bd-af9a-4b4d900db00e\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.847232 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-var-lock\") pod \"deea33e9-6b00-41bd-af9a-4b4d900db00e\" (UID: \"deea33e9-6b00-41bd-af9a-4b4d900db00e\") " Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.847269 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "deea33e9-6b00-41bd-af9a-4b4d900db00e" (UID: "deea33e9-6b00-41bd-af9a-4b4d900db00e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.847367 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-var-lock" (OuterVolumeSpecName: "var-lock") pod "deea33e9-6b00-41bd-af9a-4b4d900db00e" (UID: "deea33e9-6b00-41bd-af9a-4b4d900db00e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.847641 4492 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.847667 4492 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/deea33e9-6b00-41bd-af9a-4b4d900db00e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.853295 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deea33e9-6b00-41bd-af9a-4b4d900db00e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "deea33e9-6b00-41bd-af9a-4b4d900db00e" (UID: "deea33e9-6b00-41bd-af9a-4b4d900db00e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:44:35 crc kubenswrapper[4492]: I0220 06:44:35.948568 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/deea33e9-6b00-41bd-af9a-4b4d900db00e-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.319830 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.320777 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.321263 4492 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.321615 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.454833 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.454899 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.454916 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.454948 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.454948 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.455088 4492 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.455108 4492 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.455074 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.496559 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.497338 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321" exitCode=0 Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.497407 4492 scope.go:117] "RemoveContainer" containerID="186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.497574 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.503705 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"deea33e9-6b00-41bd-af9a-4b4d900db00e","Type":"ContainerDied","Data":"622c2ec4340bc398c79be911cfc3cbb210aa8b49a2934464537ccfb293f5fad7"} Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.503756 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622c2ec4340bc398c79be911cfc3cbb210aa8b49a2934464537ccfb293f5fad7" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.503792 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 20 06:44:36 crc kubenswrapper[4492]: E0220 06:44:36.504440 4492 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.26.25:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.510238 4492 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.510959 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.517705 4492 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.518220 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.519369 4492 scope.go:117] "RemoveContainer" containerID="28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.534949 4492 scope.go:117] "RemoveContainer" containerID="f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.547175 4492 scope.go:117] "RemoveContainer" containerID="356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.556836 4492 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.558132 4492 scope.go:117] "RemoveContainer" containerID="b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.571435 4492 scope.go:117] "RemoveContainer" containerID="cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.588783 4492 scope.go:117] "RemoveContainer" containerID="186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c" Feb 20 06:44:36 crc kubenswrapper[4492]: E0220 06:44:36.589267 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\": container with ID starting with 186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c not found: ID does not exist" containerID="186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.589339 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c"} err="failed to get container status \"186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\": rpc error: code = NotFound desc = could not find container \"186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c\": container with ID starting with 186f9211490936757a72ac36a4eade7a3868c661d49cb5e332422e7f99cdc95c not found: ID does not exist" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.589406 4492 scope.go:117] "RemoveContainer" containerID="28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac" Feb 20 06:44:36 crc kubenswrapper[4492]: E0220 06:44:36.590855 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\": container with ID starting with 28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac not found: ID does not exist" containerID="28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.590886 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac"} err="failed to get container status \"28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\": rpc error: code = NotFound desc = could not find container \"28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac\": container with ID starting with 28c25bdc755864376ffc8e4f36a96f3417e0e383f244f22c781bb56e1bcfbfac not found: ID does not exist" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.590909 4492 scope.go:117] "RemoveContainer" containerID="f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b" Feb 20 06:44:36 crc kubenswrapper[4492]: E0220 06:44:36.591319 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\": container with ID starting with f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b not found: ID does not exist" containerID="f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.591363 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b"} err="failed to get container status \"f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\": rpc error: code = NotFound desc = could not find container \"f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b\": container with ID starting with f008f6cbe5ebb9600a8606a7def53280fba888aec604f1cee676b4a5e55e8f4b not found: ID does not exist" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.591399 4492 scope.go:117] "RemoveContainer" containerID="356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c" Feb 20 06:44:36 crc kubenswrapper[4492]: E0220 06:44:36.591686 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\": container with ID starting with 356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c not found: ID does not exist" containerID="356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.591717 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c"} err="failed to get container status \"356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\": rpc error: code = NotFound desc = could not find container \"356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c\": container with ID starting with 356847f6219e220c6feb814dc41acdfe64e839f6079ac6a07f5dd067387eef0c not found: ID does not exist" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.591735 4492 scope.go:117] "RemoveContainer" containerID="b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321" Feb 20 06:44:36 crc kubenswrapper[4492]: E0220 06:44:36.592035 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\": container with ID starting with b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321 not found: ID does not exist" containerID="b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.592067 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321"} err="failed to get container status \"b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\": rpc error: code = NotFound desc = could not find container \"b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321\": container with ID starting with b75aa0cd2bee5bebb61d95372bad75d9968fcccabc0e83b12ed150024656b321 not found: ID does not exist" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.592087 4492 scope.go:117] "RemoveContainer" containerID="cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534" Feb 20 06:44:36 crc kubenswrapper[4492]: E0220 06:44:36.592348 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\": container with ID starting with cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534 not found: ID does not exist" containerID="cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534" Feb 20 06:44:36 crc kubenswrapper[4492]: I0220 06:44:36.592383 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534"} err="failed to get container status \"cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\": rpc error: code = NotFound desc = could not find container \"cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534\": container with ID starting with cce4df28bf181895c34a69ab6e26541aac51525a07fa78bb6edcf1de819f7534 not found: ID does not exist" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.260882 4492 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.261211 4492 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.261588 4492 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.261910 4492 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.262172 4492 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:37 crc kubenswrapper[4492]: I0220 06:44:37.262204 4492 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.262434 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="200ms" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.463098 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="400ms" Feb 20 06:44:37 crc kubenswrapper[4492]: I0220 06:44:37.577799 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.579926 4492 projected.go:194] Error preparing data for projected volume kube-api-access-59v8d for pod openshift-marketplace/redhat-operators-5kkhh: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.580612 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d podName:9d2d17f0-7864-41dd-8b9f-285944a57376 nodeName:}" failed. No retries permitted until 2026-02-20 06:44:41.580567902 +0000 UTC m=+238.351856870 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-59v8d" (UniqueName: "kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d") pod "redhat-operators-5kkhh" (UID: "9d2d17f0-7864-41dd-8b9f-285944a57376") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:37 crc kubenswrapper[4492]: I0220 06:44:37.587194 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 20 06:44:37 crc kubenswrapper[4492]: E0220 06:44:37.864269 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="800ms" Feb 20 06:44:38 crc kubenswrapper[4492]: E0220 06:44:38.665711 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="1.6s" Feb 20 06:44:39 crc kubenswrapper[4492]: E0220 06:44:39.998543 4492 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 192.168.26.25:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-operators-5kkhh.1895e16a0b0441a0 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-operators-5kkhh,UID:9d2d17f0-7864-41dd-8b9f-285944a57376,APIVersion:v1,ResourceVersion:29621,FieldPath:,},Reason:FailedMount,Message:MountVolume.SetUp failed for volume \"kube-api-access-59v8d\" : failed to fetch token: Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token\": dial tcp 192.168.26.25:6443: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-20 06:44:34.007769504 +0000 UTC m=+230.779058482,LastTimestamp:2026-02-20 06:44:34.007769504 +0000 UTC m=+230.779058482,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 20 06:44:40 crc kubenswrapper[4492]: E0220 06:44:40.267273 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="3.2s" Feb 20 06:44:41 crc kubenswrapper[4492]: I0220 06:44:41.621998 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:41 crc kubenswrapper[4492]: E0220 06:44:41.622857 4492 projected.go:194] Error preparing data for projected volume kube-api-access-59v8d for pod openshift-marketplace/redhat-operators-5kkhh: failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:41 crc kubenswrapper[4492]: E0220 06:44:41.622986 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d podName:9d2d17f0-7864-41dd-8b9f-285944a57376 nodeName:}" failed. No retries permitted until 2026-02-20 06:44:49.622952395 +0000 UTC m=+246.394241373 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-59v8d" (UniqueName: "kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d") pod "redhat-operators-5kkhh" (UID: "9d2d17f0-7864-41dd-8b9f-285944a57376") : failed to fetch token: Post "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/redhat-operators/token": dial tcp 192.168.26.25:6443: connect: connection refused Feb 20 06:44:43 crc kubenswrapper[4492]: E0220 06:44:43.467747 4492 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.25:6443: connect: connection refused" interval="6.4s" Feb 20 06:44:43 crc kubenswrapper[4492]: I0220 06:44:43.563379 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:44 crc kubenswrapper[4492]: E0220 06:44:44.560993 4492 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 192.168.26.25:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" volumeName="registry-storage" Feb 20 06:44:45 crc kubenswrapper[4492]: I0220 06:44:45.556538 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:45 crc kubenswrapper[4492]: I0220 06:44:45.557863 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:45 crc kubenswrapper[4492]: I0220 06:44:45.572785 4492 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:45 crc kubenswrapper[4492]: I0220 06:44:45.572823 4492 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:45 crc kubenswrapper[4492]: E0220 06:44:45.573110 4492 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:45 crc kubenswrapper[4492]: I0220 06:44:45.573633 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:45 crc kubenswrapper[4492]: W0220 06:44:45.593499 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-c11ad3703adab866dfb3b493f294c5e6392bac732a2710a9340cc599831e11b3 WatchSource:0}: Error finding container c11ad3703adab866dfb3b493f294c5e6392bac732a2710a9340cc599831e11b3: Status 404 returned error can't find the container with id c11ad3703adab866dfb3b493f294c5e6392bac732a2710a9340cc599831e11b3 Feb 20 06:44:46 crc kubenswrapper[4492]: I0220 06:44:46.562559 4492 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="72a60ce351fcbe750d6ee412a5af3208c2d32c563e736bb3616b3462e6d41090" exitCode=0 Feb 20 06:44:46 crc kubenswrapper[4492]: I0220 06:44:46.562644 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"72a60ce351fcbe750d6ee412a5af3208c2d32c563e736bb3616b3462e6d41090"} Feb 20 06:44:46 crc kubenswrapper[4492]: I0220 06:44:46.562894 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c11ad3703adab866dfb3b493f294c5e6392bac732a2710a9340cc599831e11b3"} Feb 20 06:44:46 crc kubenswrapper[4492]: I0220 06:44:46.563203 4492 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:46 crc kubenswrapper[4492]: I0220 06:44:46.563220 4492 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:46 crc kubenswrapper[4492]: E0220 06:44:46.563549 4492 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:46 crc kubenswrapper[4492]: I0220 06:44:46.563719 4492 status_manager.go:851] "Failed to get status for pod" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.25:6443: connect: connection refused" Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.576837 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4b9e21acbb21fe90f7ce144f91e15d0657739c413a2957d51513022e1491ac6c"} Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.577149 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cbbcd972bc1d4cd863a5b244b2d02857367bd6b8dd907140d7886459f957bb03"} Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.577163 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2cd0636921f3895443378cf58b1e68d99f90b589f408c0d53ee2e0730d41c800"} Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.577174 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5e09f3914f068a5a6c9127c7116a674836b21146e94b648614f7ef6ed7f79afa"} Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.577185 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fc3726f208753a1037860a0d6de63ae041e99a1be0819ff649faf89665e7a4ff"} Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.577420 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.577568 4492 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:47 crc kubenswrapper[4492]: I0220 06:44:47.577587 4492 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:48 crc kubenswrapper[4492]: I0220 06:44:48.556423 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:48 crc kubenswrapper[4492]: I0220 06:44:48.556716 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:44:48 crc kubenswrapper[4492]: I0220 06:44:48.583026 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 06:44:48 crc kubenswrapper[4492]: I0220 06:44:48.583056 4492 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1" exitCode=1 Feb 20 06:44:48 crc kubenswrapper[4492]: I0220 06:44:48.583082 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1"} Feb 20 06:44:48 crc kubenswrapper[4492]: I0220 06:44:48.583368 4492 scope.go:117] "RemoveContainer" containerID="416cc49ba4ba78251db7490e7e39fde6756545e9d48dd4ef33516052ecd140a1" Feb 20 06:44:49 crc kubenswrapper[4492]: I0220 06:44:49.590597 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 20 06:44:49 crc kubenswrapper[4492]: I0220 06:44:49.590902 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b7c6ab2e67802f4c333f91c2a58d072af5b1272ad0e1913dcd5c9f62b9c5ee23"} Feb 20 06:44:49 crc kubenswrapper[4492]: I0220 06:44:49.722955 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:49 crc kubenswrapper[4492]: I0220 06:44:49.741782 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59v8d\" (UniqueName: \"kubernetes.io/projected/9d2d17f0-7864-41dd-8b9f-285944a57376-kube-api-access-59v8d\") pod \"redhat-operators-5kkhh\" (UID: \"9d2d17f0-7864-41dd-8b9f-285944a57376\") " pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.028041 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.036943 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.574437 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.574683 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.579914 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.597590 4492 generic.go:334] "Generic (PLEG): container finished" podID="9d2d17f0-7864-41dd-8b9f-285944a57376" containerID="4f5ed915fb9d190c617b06f1d9c78be976694620e7794052c30a8cb158697a6e" exitCode=0 Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.597631 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kkhh" event={"ID":"9d2d17f0-7864-41dd-8b9f-285944a57376","Type":"ContainerDied","Data":"4f5ed915fb9d190c617b06f1d9c78be976694620e7794052c30a8cb158697a6e"} Feb 20 06:44:50 crc kubenswrapper[4492]: I0220 06:44:50.597661 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kkhh" event={"ID":"9d2d17f0-7864-41dd-8b9f-285944a57376","Type":"ContainerStarted","Data":"b1117131d8c2bf34e2bf85944179fa7e6dc5e2cc58a193cab0869a9377987e33"} Feb 20 06:44:51 crc kubenswrapper[4492]: I0220 06:44:51.605506 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kkhh" event={"ID":"9d2d17f0-7864-41dd-8b9f-285944a57376","Type":"ContainerStarted","Data":"a8ded16123d19c23162f9449a7d3ece067f68cd5d5d7a5bd8c5dffcd6ec186e1"} Feb 20 06:44:52 crc kubenswrapper[4492]: I0220 06:44:52.615092 4492 generic.go:334] "Generic (PLEG): container finished" podID="9d2d17f0-7864-41dd-8b9f-285944a57376" containerID="a8ded16123d19c23162f9449a7d3ece067f68cd5d5d7a5bd8c5dffcd6ec186e1" exitCode=0 Feb 20 06:44:52 crc kubenswrapper[4492]: I0220 06:44:52.615155 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kkhh" event={"ID":"9d2d17f0-7864-41dd-8b9f-285944a57376","Type":"ContainerDied","Data":"a8ded16123d19c23162f9449a7d3ece067f68cd5d5d7a5bd8c5dffcd6ec186e1"} Feb 20 06:44:52 crc kubenswrapper[4492]: I0220 06:44:52.966209 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:44:53 crc kubenswrapper[4492]: I0220 06:44:53.053458 4492 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:44:53 crc kubenswrapper[4492]: I0220 06:44:53.569028 4492 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7590f1c3-3680-4e8a-85c9-e0432697f430" Feb 20 06:44:53 crc kubenswrapper[4492]: I0220 06:44:53.621782 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5kkhh" event={"ID":"9d2d17f0-7864-41dd-8b9f-285944a57376","Type":"ContainerStarted","Data":"fd87422395c55e5d78899d2f6ad2b2e737505225981734b7df193596a3fdc881"} Feb 20 06:44:53 crc kubenswrapper[4492]: I0220 06:44:53.622189 4492 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:53 crc kubenswrapper[4492]: I0220 06:44:53.622211 4492 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="33d8f5bf-8c00-4260-9e5b-e188570d4715" Feb 20 06:44:53 crc kubenswrapper[4492]: I0220 06:44:53.646109 4492 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7590f1c3-3680-4e8a-85c9-e0432697f430" Feb 20 06:44:53 crc kubenswrapper[4492]: W0220 06:44:53.778244 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bcd6b9a_5837_48eb_98d0_691541bfff4d.slice/crio-90222931a91bc045718c4b1e5edf80f8255657aaf6d21f43c7a29895550d3a4c WatchSource:0}: Error finding container 90222931a91bc045718c4b1e5edf80f8255657aaf6d21f43c7a29895550d3a4c: Status 404 returned error can't find the container with id 90222931a91bc045718c4b1e5edf80f8255657aaf6d21f43c7a29895550d3a4c Feb 20 06:44:54 crc kubenswrapper[4492]: I0220 06:44:54.627076 4492 generic.go:334] "Generic (PLEG): container finished" podID="2bcd6b9a-5837-48eb-98d0-691541bfff4d" containerID="9b2b0275e3f1114c26e011111d6639d013e530017f63d98b7d9f74aa23995022" exitCode=0 Feb 20 06:44:54 crc kubenswrapper[4492]: I0220 06:44:54.627151 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cms44" event={"ID":"2bcd6b9a-5837-48eb-98d0-691541bfff4d","Type":"ContainerDied","Data":"9b2b0275e3f1114c26e011111d6639d013e530017f63d98b7d9f74aa23995022"} Feb 20 06:44:54 crc kubenswrapper[4492]: I0220 06:44:54.627833 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cms44" event={"ID":"2bcd6b9a-5837-48eb-98d0-691541bfff4d","Type":"ContainerStarted","Data":"90222931a91bc045718c4b1e5edf80f8255657aaf6d21f43c7a29895550d3a4c"} Feb 20 06:44:54 crc kubenswrapper[4492]: I0220 06:44:54.742937 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:44:54 crc kubenswrapper[4492]: I0220 06:44:54.748331 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:44:55 crc kubenswrapper[4492]: I0220 06:44:55.634482 4492 generic.go:334] "Generic (PLEG): container finished" podID="2bcd6b9a-5837-48eb-98d0-691541bfff4d" containerID="2489ee9d4500003c4768540c845661e14ecfd19a62f8b508a76333b3bbf28d08" exitCode=0 Feb 20 06:44:55 crc kubenswrapper[4492]: I0220 06:44:55.635379 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cms44" event={"ID":"2bcd6b9a-5837-48eb-98d0-691541bfff4d","Type":"ContainerDied","Data":"2489ee9d4500003c4768540c845661e14ecfd19a62f8b508a76333b3bbf28d08"} Feb 20 06:44:56 crc kubenswrapper[4492]: I0220 06:44:56.641549 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cms44" event={"ID":"2bcd6b9a-5837-48eb-98d0-691541bfff4d","Type":"ContainerStarted","Data":"eb52b7a9111bce5dbe379446f1a1b036ca35cb5b6af29b58cb8692cb2e049be7"} Feb 20 06:45:00 crc kubenswrapper[4492]: I0220 06:45:00.037354 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:45:00 crc kubenswrapper[4492]: I0220 06:45:00.040119 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:45:00 crc kubenswrapper[4492]: I0220 06:45:00.077944 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:45:00 crc kubenswrapper[4492]: I0220 06:45:00.711857 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5kkhh" Feb 20 06:45:02 crc kubenswrapper[4492]: I0220 06:45:02.968804 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 20 06:45:03 crc kubenswrapper[4492]: I0220 06:45:03.531699 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 20 06:45:03 crc kubenswrapper[4492]: I0220 06:45:03.932003 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:45:03 crc kubenswrapper[4492]: I0220 06:45:03.932066 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:45:03 crc kubenswrapper[4492]: I0220 06:45:03.970602 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:45:04 crc kubenswrapper[4492]: I0220 06:45:04.126158 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 20 06:45:04 crc kubenswrapper[4492]: I0220 06:45:04.603045 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 20 06:45:04 crc kubenswrapper[4492]: I0220 06:45:04.728739 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cms44" Feb 20 06:45:04 crc kubenswrapper[4492]: I0220 06:45:04.738233 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 20 06:45:04 crc kubenswrapper[4492]: I0220 06:45:04.739780 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 20 06:45:04 crc kubenswrapper[4492]: I0220 06:45:04.905780 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.328658 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.388720 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.439657 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.445984 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.619378 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.741086 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.789278 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 20 06:45:05 crc kubenswrapper[4492]: I0220 06:45:05.825103 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.046785 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.056025 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.472668 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.481739 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.533306 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.561291 4492 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.631081 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.662430 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 20 06:45:06 crc kubenswrapper[4492]: I0220 06:45:06.878755 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 20 06:45:07 crc kubenswrapper[4492]: I0220 06:45:07.005764 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 20 06:45:07 crc kubenswrapper[4492]: I0220 06:45:07.188187 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 20 06:45:07 crc kubenswrapper[4492]: I0220 06:45:07.381581 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 06:45:07 crc kubenswrapper[4492]: I0220 06:45:07.458217 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 20 06:45:07 crc kubenswrapper[4492]: I0220 06:45:07.509318 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.014903 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.313000 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.315241 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.379738 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.533147 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.562369 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.580120 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.632803 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.708345 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.840003 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 20 06:45:08 crc kubenswrapper[4492]: I0220 06:45:08.910006 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.074341 4492 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.240369 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.284544 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.310927 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.389896 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.446633 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.461314 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.464449 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.497667 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.515428 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.516862 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.554883 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.568506 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.569908 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.636663 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.732958 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.796691 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.821543 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.850616 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.898927 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.925809 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.970802 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.983950 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 20 06:45:09 crc kubenswrapper[4492]: I0220 06:45:09.987071 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.222854 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.241192 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.274664 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.384907 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.392168 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.503004 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.523152 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.535608 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.630280 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.668140 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.708698 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.760809 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.773247 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.798118 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.831114 4492 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.845274 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.907737 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 20 06:45:10 crc kubenswrapper[4492]: I0220 06:45:10.915860 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.095191 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.402217 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.810612 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.811824 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.826248 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.844361 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.897493 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.916902 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.949009 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 20 06:45:11 crc kubenswrapper[4492]: I0220 06:45:11.956648 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.025767 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.062916 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.063992 4492 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.065495 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cms44" podStartSLOduration=37.552173314 podStartE2EDuration="39.065458722s" podCreationTimestamp="2026-02-20 06:44:33 +0000 UTC" firstStartedPulling="2026-02-20 06:44:54.629041786 +0000 UTC m=+251.400330765" lastFinishedPulling="2026-02-20 06:44:56.142327196 +0000 UTC m=+252.913616173" observedRunningTime="2026-02-20 06:44:56.656903016 +0000 UTC m=+253.428191995" watchObservedRunningTime="2026-02-20 06:45:12.065458722 +0000 UTC m=+268.836747699" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.066223 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5kkhh" podStartSLOduration=36.496410675 podStartE2EDuration="39.066217652s" podCreationTimestamp="2026-02-20 06:44:33 +0000 UTC" firstStartedPulling="2026-02-20 06:44:50.599092447 +0000 UTC m=+247.370381425" lastFinishedPulling="2026-02-20 06:44:53.168899424 +0000 UTC m=+249.940188402" observedRunningTime="2026-02-20 06:44:53.644059095 +0000 UTC m=+250.415348073" watchObservedRunningTime="2026-02-20 06:45:12.066217652 +0000 UTC m=+268.837506630" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.067616 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.067656 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.067672 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cms44","openshift-marketplace/redhat-operators-5kkhh"] Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.071174 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.071505 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.079286 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.124501 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.124450802 podStartE2EDuration="19.124450802s" podCreationTimestamp="2026-02-20 06:44:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:45:12.121913101 +0000 UTC m=+268.893202078" watchObservedRunningTime="2026-02-20 06:45:12.124450802 +0000 UTC m=+268.895739770" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.136692 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.158462 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.177717 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.194216 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.228920 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.255502 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.326834 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.456766 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.534379 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.595112 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.596275 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.633627 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.719657 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.722674 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.835824 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.843945 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.848773 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.881036 4492 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 20 06:45:12 crc kubenswrapper[4492]: I0220 06:45:12.975041 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.041974 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.059917 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.103200 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.105501 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.125050 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.164368 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.195240 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.238102 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.240357 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.245517 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.246987 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.276923 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.396742 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.414664 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.414665 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.527518 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.582705 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.626373 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.642592 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.660044 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.697035 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.831224 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.848950 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.891823 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.902937 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.977038 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 20 06:45:13 crc kubenswrapper[4492]: I0220 06:45:13.994046 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.039083 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.088366 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.088895 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.177628 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.228214 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.235949 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.249682 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.252748 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.323165 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.439158 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.469511 4492 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.469740 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://73c73a886d5369936df5a17ed652bcecc0dc75b8c03c83adaa1b6082e8f07c82" gracePeriod=5 Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.552244 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.577179 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.588787 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.600036 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.612365 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.627757 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.656909 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.686802 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.696960 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.760404 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.842733 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 20 06:45:14 crc kubenswrapper[4492]: I0220 06:45:14.896790 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.124659 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.182859 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.395418 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.541674 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.546368 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.556782 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.572052 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.594133 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.620185 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.670048 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.811571 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 20 06:45:15 crc kubenswrapper[4492]: I0220 06:45:15.997666 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.000642 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.010770 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.115914 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.283769 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.317974 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.388462 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.428555 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.446581 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.526894 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.527941 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.599970 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.616397 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.656300 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.720666 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.730638 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.735785 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.814086 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.826007 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.876958 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.877811 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.904015 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.915615 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.977089 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.989130 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 06:45:16 crc kubenswrapper[4492]: I0220 06:45:16.992277 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.202880 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.237441 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv"] Feb 20 06:45:17 crc kubenswrapper[4492]: E0220 06:45:17.237701 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" containerName="installer" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.237719 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" containerName="installer" Feb 20 06:45:17 crc kubenswrapper[4492]: E0220 06:45:17.237741 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.237747 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.237831 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.237846 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="deea33e9-6b00-41bd-af9a-4b4d900db00e" containerName="installer" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.238194 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.240465 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.240630 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.246864 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.253221 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv"] Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.273220 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-secret-volume\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.273279 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npfh8\" (UniqueName: \"kubernetes.io/projected/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-kube-api-access-npfh8\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.273323 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-config-volume\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.337119 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.374815 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-secret-volume\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.374855 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npfh8\" (UniqueName: \"kubernetes.io/projected/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-kube-api-access-npfh8\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.374884 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-config-volume\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.375750 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-config-volume\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.397171 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-secret-volume\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.400295 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npfh8\" (UniqueName: \"kubernetes.io/projected/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-kube-api-access-npfh8\") pod \"collect-profiles-29526165-7vskv\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.446039 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.477150 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.478082 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.498262 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.554678 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.626106 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.695077 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.704466 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.732089 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv"] Feb 20 06:45:17 crc kubenswrapper[4492]: W0220 06:45:17.738083 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod284c2ed2_e161_4e3b_9ceb_891c71dc4e1f.slice/crio-3fa9a54b9c663681f45e026ee40d8b53699d368cb00141dba6411e5990cb35c2 WatchSource:0}: Error finding container 3fa9a54b9c663681f45e026ee40d8b53699d368cb00141dba6411e5990cb35c2: Status 404 returned error can't find the container with id 3fa9a54b9c663681f45e026ee40d8b53699d368cb00141dba6411e5990cb35c2 Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.766328 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" event={"ID":"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f","Type":"ContainerStarted","Data":"3fa9a54b9c663681f45e026ee40d8b53699d368cb00141dba6411e5990cb35c2"} Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.771367 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.798663 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.806195 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.812795 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.853465 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.982008 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 20 06:45:17 crc kubenswrapper[4492]: I0220 06:45:17.988624 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.465225 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.520943 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.567390 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.601940 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.718815 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.771657 4492 generic.go:334] "Generic (PLEG): container finished" podID="284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" containerID="db8b5a49afb523ad43b8c4093110a8b61adbac549e7df2fe0a92fa35ebeef35a" exitCode=0 Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.771755 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" event={"ID":"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f","Type":"ContainerDied","Data":"db8b5a49afb523ad43b8c4093110a8b61adbac549e7df2fe0a92fa35ebeef35a"} Feb 20 06:45:18 crc kubenswrapper[4492]: I0220 06:45:18.905754 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.187302 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.201525 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.298262 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.386348 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.521269 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.685741 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.722499 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.746773 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.752079 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.754515 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.777884 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.777930 4492 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="73c73a886d5369936df5a17ed652bcecc0dc75b8c03c83adaa1b6082e8f07c82" exitCode=137 Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.851629 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.937419 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.953638 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.981399 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.991811 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 20 06:45:19 crc kubenswrapper[4492]: I0220 06:45:19.996151 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.006369 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npfh8\" (UniqueName: \"kubernetes.io/projected/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-kube-api-access-npfh8\") pod \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.006434 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-config-volume\") pod \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.006461 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-secret-volume\") pod \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\" (UID: \"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.007903 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-config-volume" (OuterVolumeSpecName: "config-volume") pod "284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" (UID: "284c2ed2-e161-4e3b-9ceb-891c71dc4e1f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.011296 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" (UID: "284c2ed2-e161-4e3b-9ceb-891c71dc4e1f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.011529 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-kube-api-access-npfh8" (OuterVolumeSpecName: "kube-api-access-npfh8") pod "284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" (UID: "284c2ed2-e161-4e3b-9ceb-891c71dc4e1f"). InnerVolumeSpecName "kube-api-access-npfh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.035162 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.035303 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.102053 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.107835 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.107894 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.107943 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108003 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108025 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108211 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108208 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108251 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108260 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108276 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npfh8\" (UniqueName: \"kubernetes.io/projected/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-kube-api-access-npfh8\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108281 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.108286 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.112921 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.172129 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.209841 4492 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.209866 4492 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.209878 4492 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.209891 4492 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.209899 4492 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.265514 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.447757 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.635815 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.783908 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.783908 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv" event={"ID":"284c2ed2-e161-4e3b-9ceb-891c71dc4e1f","Type":"ContainerDied","Data":"3fa9a54b9c663681f45e026ee40d8b53699d368cb00141dba6411e5990cb35c2"} Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.784060 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fa9a54b9c663681f45e026ee40d8b53699d368cb00141dba6411e5990cb35c2" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.785601 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.785676 4492 scope.go:117] "RemoveContainer" containerID="73c73a886d5369936df5a17ed652bcecc0dc75b8c03c83adaa1b6082e8f07c82" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.785737 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 20 06:45:20 crc kubenswrapper[4492]: I0220 06:45:20.857267 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 20 06:45:21 crc kubenswrapper[4492]: I0220 06:45:21.094003 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 20 06:45:21 crc kubenswrapper[4492]: I0220 06:45:21.562734 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 20 06:45:21 crc kubenswrapper[4492]: I0220 06:45:21.781079 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 20 06:45:22 crc kubenswrapper[4492]: I0220 06:45:22.003908 4492 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 20 06:45:22 crc kubenswrapper[4492]: I0220 06:45:22.335353 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 20 06:45:22 crc kubenswrapper[4492]: I0220 06:45:22.854683 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.747728 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8fwhv"] Feb 20 06:45:35 crc kubenswrapper[4492]: E0220 06:45:35.748331 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" containerName="collect-profiles" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.748346 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" containerName="collect-profiles" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.748443 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" containerName="collect-profiles" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.749125 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.750850 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.754230 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8fwhv"] Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.758157 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-utilities\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.758215 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-catalog-content\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.758248 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7dkv\" (UniqueName: \"kubernetes.io/projected/c83fd1dd-24ef-45f8-a3ff-264d74862d94-kube-api-access-b7dkv\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.858937 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-utilities\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.859018 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-catalog-content\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.859069 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7dkv\" (UniqueName: \"kubernetes.io/projected/c83fd1dd-24ef-45f8-a3ff-264d74862d94-kube-api-access-b7dkv\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.859308 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-utilities\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.859362 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-catalog-content\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.875594 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7dkv\" (UniqueName: \"kubernetes.io/projected/c83fd1dd-24ef-45f8-a3ff-264d74862d94-kube-api-access-b7dkv\") pod \"community-operators-8fwhv\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.945903 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vc5ws"] Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.946957 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.952139 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.955817 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vc5ws"] Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.959793 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgxzk\" (UniqueName: \"kubernetes.io/projected/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-kube-api-access-qgxzk\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.959851 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-utilities\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:35 crc kubenswrapper[4492]: I0220 06:45:35.959884 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-catalog-content\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.060760 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgxzk\" (UniqueName: \"kubernetes.io/projected/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-kube-api-access-qgxzk\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.060827 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-utilities\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.060867 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-catalog-content\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.061319 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-utilities\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.061332 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-catalog-content\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.062420 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.075907 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgxzk\" (UniqueName: \"kubernetes.io/projected/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-kube-api-access-qgxzk\") pod \"certified-operators-vc5ws\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.198379 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8fwhv"] Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.260077 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.353322 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7ggr"] Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.353623 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" podUID="8f9c533a-3aa4-4413-914d-6d24f234c092" containerName="controller-manager" containerID="cri-o://e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe" gracePeriod=30 Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.364375 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs"] Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.365004 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" podUID="80f87acb-f24f-4f79-9677-227c929e0497" containerName="route-controller-manager" containerID="cri-o://7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc" gracePeriod=30 Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.473799 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vc5ws"] Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.652574 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.655087 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676409 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-config\") pod \"8f9c533a-3aa4-4413-914d-6d24f234c092\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676459 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-config\") pod \"80f87acb-f24f-4f79-9677-227c929e0497\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676514 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-client-ca\") pod \"80f87acb-f24f-4f79-9677-227c929e0497\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676548 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-client-ca\") pod \"8f9c533a-3aa4-4413-914d-6d24f234c092\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676571 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-proxy-ca-bundles\") pod \"8f9c533a-3aa4-4413-914d-6d24f234c092\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676598 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80f87acb-f24f-4f79-9677-227c929e0497-serving-cert\") pod \"80f87acb-f24f-4f79-9677-227c929e0497\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676618 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6twn4\" (UniqueName: \"kubernetes.io/projected/80f87acb-f24f-4f79-9677-227c929e0497-kube-api-access-6twn4\") pod \"80f87acb-f24f-4f79-9677-227c929e0497\" (UID: \"80f87acb-f24f-4f79-9677-227c929e0497\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.676639 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tx8r2\" (UniqueName: \"kubernetes.io/projected/8f9c533a-3aa4-4413-914d-6d24f234c092-kube-api-access-tx8r2\") pod \"8f9c533a-3aa4-4413-914d-6d24f234c092\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.677447 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-config" (OuterVolumeSpecName: "config") pod "8f9c533a-3aa4-4413-914d-6d24f234c092" (UID: "8f9c533a-3aa4-4413-914d-6d24f234c092"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.678051 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-client-ca" (OuterVolumeSpecName: "client-ca") pod "8f9c533a-3aa4-4413-914d-6d24f234c092" (UID: "8f9c533a-3aa4-4413-914d-6d24f234c092"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.678982 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-config" (OuterVolumeSpecName: "config") pod "80f87acb-f24f-4f79-9677-227c929e0497" (UID: "80f87acb-f24f-4f79-9677-227c929e0497"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.679438 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-client-ca" (OuterVolumeSpecName: "client-ca") pod "80f87acb-f24f-4f79-9677-227c929e0497" (UID: "80f87acb-f24f-4f79-9677-227c929e0497"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.679841 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8f9c533a-3aa4-4413-914d-6d24f234c092" (UID: "8f9c533a-3aa4-4413-914d-6d24f234c092"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.684881 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9c533a-3aa4-4413-914d-6d24f234c092-kube-api-access-tx8r2" (OuterVolumeSpecName: "kube-api-access-tx8r2") pod "8f9c533a-3aa4-4413-914d-6d24f234c092" (UID: "8f9c533a-3aa4-4413-914d-6d24f234c092"). InnerVolumeSpecName "kube-api-access-tx8r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.690809 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80f87acb-f24f-4f79-9677-227c929e0497-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "80f87acb-f24f-4f79-9677-227c929e0497" (UID: "80f87acb-f24f-4f79-9677-227c929e0497"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.692526 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80f87acb-f24f-4f79-9677-227c929e0497-kube-api-access-6twn4" (OuterVolumeSpecName: "kube-api-access-6twn4") pod "80f87acb-f24f-4f79-9677-227c929e0497" (UID: "80f87acb-f24f-4f79-9677-227c929e0497"). InnerVolumeSpecName "kube-api-access-6twn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.777801 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f9c533a-3aa4-4413-914d-6d24f234c092-serving-cert\") pod \"8f9c533a-3aa4-4413-914d-6d24f234c092\" (UID: \"8f9c533a-3aa4-4413-914d-6d24f234c092\") " Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778120 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778134 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778147 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80f87acb-f24f-4f79-9677-227c929e0497-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778158 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778166 4492 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8f9c533a-3aa4-4413-914d-6d24f234c092-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778175 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80f87acb-f24f-4f79-9677-227c929e0497-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778184 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6twn4\" (UniqueName: \"kubernetes.io/projected/80f87acb-f24f-4f79-9677-227c929e0497-kube-api-access-6twn4\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.778193 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tx8r2\" (UniqueName: \"kubernetes.io/projected/8f9c533a-3aa4-4413-914d-6d24f234c092-kube-api-access-tx8r2\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.780254 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9c533a-3aa4-4413-914d-6d24f234c092-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8f9c533a-3aa4-4413-914d-6d24f234c092" (UID: "8f9c533a-3aa4-4413-914d-6d24f234c092"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.855625 4492 generic.go:334] "Generic (PLEG): container finished" podID="80f87acb-f24f-4f79-9677-227c929e0497" containerID="7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc" exitCode=0 Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.855712 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" event={"ID":"80f87acb-f24f-4f79-9677-227c929e0497","Type":"ContainerDied","Data":"7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.855743 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" event={"ID":"80f87acb-f24f-4f79-9677-227c929e0497","Type":"ContainerDied","Data":"986edab851de5f00df0a850fdcf4fce712f5ba91fe01b7de1075f974ca9e8517"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.855744 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.855765 4492 scope.go:117] "RemoveContainer" containerID="7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.857537 4492 generic.go:334] "Generic (PLEG): container finished" podID="8f9c533a-3aa4-4413-914d-6d24f234c092" containerID="e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe" exitCode=0 Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.857618 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" event={"ID":"8f9c533a-3aa4-4413-914d-6d24f234c092","Type":"ContainerDied","Data":"e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.857651 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" event={"ID":"8f9c533a-3aa4-4413-914d-6d24f234c092","Type":"ContainerDied","Data":"f8f98e3bb00b3330ecc2d6b21ed8cc693311c4df7e4118178e78e24e75f35741"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.857730 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v7ggr" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.862114 4492 generic.go:334] "Generic (PLEG): container finished" podID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerID="91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68" exitCode=0 Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.862219 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vc5ws" event={"ID":"b0e828a1-a902-411f-8b55-f12b3a8c3e1c","Type":"ContainerDied","Data":"91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.862260 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vc5ws" event={"ID":"b0e828a1-a902-411f-8b55-f12b3a8c3e1c","Type":"ContainerStarted","Data":"ae375b830ddb6e16c922c8f93a22e0a1528e077c01fafc00fb609da80c36acbf"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.865834 4492 generic.go:334] "Generic (PLEG): container finished" podID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerID="2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c" exitCode=0 Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.865882 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fwhv" event={"ID":"c83fd1dd-24ef-45f8-a3ff-264d74862d94","Type":"ContainerDied","Data":"2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.865908 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fwhv" event={"ID":"c83fd1dd-24ef-45f8-a3ff-264d74862d94","Type":"ContainerStarted","Data":"565a9d7c3c86b616208861d203cd85842e261bf3822b88ae16df4d9653bebb95"} Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.873403 4492 scope.go:117] "RemoveContainer" containerID="7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc" Feb 20 06:45:36 crc kubenswrapper[4492]: E0220 06:45:36.873897 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc\": container with ID starting with 7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc not found: ID does not exist" containerID="7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.873928 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc"} err="failed to get container status \"7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc\": rpc error: code = NotFound desc = could not find container \"7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc\": container with ID starting with 7c02efa137dd44e70d72f31259deb135ae9247a5ec7d13eb51ca07dd773bf1fc not found: ID does not exist" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.873953 4492 scope.go:117] "RemoveContainer" containerID="e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.879924 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f9c533a-3aa4-4413-914d-6d24f234c092-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.896925 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7ggr"] Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.900738 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v7ggr"] Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.912003 4492 scope.go:117] "RemoveContainer" containerID="e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe" Feb 20 06:45:36 crc kubenswrapper[4492]: E0220 06:45:36.912602 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe\": container with ID starting with e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe not found: ID does not exist" containerID="e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.912631 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe"} err="failed to get container status \"e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe\": rpc error: code = NotFound desc = could not find container \"e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe\": container with ID starting with e7d1fc5be87ee004a26d422e63b6c3b3baf7257b556cd26702a9a770868b7cbe not found: ID does not exist" Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.913814 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs"] Feb 20 06:45:36 crc kubenswrapper[4492]: I0220 06:45:36.916436 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v2kxs"] Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.566156 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80f87acb-f24f-4f79-9677-227c929e0497" path="/var/lib/kubelet/pods/80f87acb-f24f-4f79-9677-227c929e0497/volumes" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.566690 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9c533a-3aa4-4413-914d-6d24f234c092" path="/var/lib/kubelet/pods/8f9c533a-3aa4-4413-914d-6d24f234c092/volumes" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.601667 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls"] Feb 20 06:45:37 crc kubenswrapper[4492]: E0220 06:45:37.601974 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80f87acb-f24f-4f79-9677-227c929e0497" containerName="route-controller-manager" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.601990 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="80f87acb-f24f-4f79-9677-227c929e0497" containerName="route-controller-manager" Feb 20 06:45:37 crc kubenswrapper[4492]: E0220 06:45:37.602008 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9c533a-3aa4-4413-914d-6d24f234c092" containerName="controller-manager" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.602016 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9c533a-3aa4-4413-914d-6d24f234c092" containerName="controller-manager" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.602119 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9c533a-3aa4-4413-914d-6d24f234c092" containerName="controller-manager" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.602134 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="80f87acb-f24f-4f79-9677-227c929e0497" containerName="route-controller-manager" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.602582 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.605989 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.606140 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.606371 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.606525 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.606609 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d675855ff-hqhh5"] Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.607077 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.607362 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.612109 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.612733 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.612871 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.612984 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.613174 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.613328 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.613890 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls"] Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.616426 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d675855ff-hqhh5"] Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.619164 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.625675 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.692887 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-client-ca\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.692972 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c36f41-337f-4a4d-b0ec-3e45091fbef2-serving-cert\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.693039 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-config\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.693092 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-serving-cert\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.693135 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdzn4\" (UniqueName: \"kubernetes.io/projected/32c36f41-337f-4a4d-b0ec-3e45091fbef2-kube-api-access-sdzn4\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.693166 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-client-ca\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.693385 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltd6h\" (UniqueName: \"kubernetes.io/projected/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-kube-api-access-ltd6h\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.693451 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-config\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.693534 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-proxy-ca-bundles\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.794111 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-proxy-ca-bundles\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.794178 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-client-ca\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.794205 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c36f41-337f-4a4d-b0ec-3e45091fbef2-serving-cert\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.794224 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-config\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.794255 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-serving-cert\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.794278 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdzn4\" (UniqueName: \"kubernetes.io/projected/32c36f41-337f-4a4d-b0ec-3e45091fbef2-kube-api-access-sdzn4\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.795271 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-client-ca\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.795317 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltd6h\" (UniqueName: \"kubernetes.io/projected/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-kube-api-access-ltd6h\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.795341 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-config\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.795404 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-client-ca\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.795467 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-proxy-ca-bundles\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.795554 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-config\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.796052 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-client-ca\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.796501 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-config\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.799140 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c36f41-337f-4a4d-b0ec-3e45091fbef2-serving-cert\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.799158 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-serving-cert\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.809936 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltd6h\" (UniqueName: \"kubernetes.io/projected/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-kube-api-access-ltd6h\") pod \"controller-manager-6d675855ff-hqhh5\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.810485 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdzn4\" (UniqueName: \"kubernetes.io/projected/32c36f41-337f-4a4d-b0ec-3e45091fbef2-kube-api-access-sdzn4\") pod \"route-controller-manager-5c6c76fcf7-6b2ls\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.875285 4492 generic.go:334] "Generic (PLEG): container finished" podID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerID="c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e" exitCode=0 Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.875351 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vc5ws" event={"ID":"b0e828a1-a902-411f-8b55-f12b3a8c3e1c","Type":"ContainerDied","Data":"c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e"} Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.879868 4492 generic.go:334] "Generic (PLEG): container finished" podID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerID="5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9" exitCode=0 Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.879916 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fwhv" event={"ID":"c83fd1dd-24ef-45f8-a3ff-264d74862d94","Type":"ContainerDied","Data":"5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9"} Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.924793 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:37 crc kubenswrapper[4492]: I0220 06:45:37.935994 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.126145 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls"] Feb 20 06:45:38 crc kubenswrapper[4492]: W0220 06:45:38.144550 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32c36f41_337f_4a4d_b0ec_3e45091fbef2.slice/crio-1984cb08147761de40eee4fdcf3bcfa712e8bf8a7d1292247473d243de6b23d6 WatchSource:0}: Error finding container 1984cb08147761de40eee4fdcf3bcfa712e8bf8a7d1292247473d243de6b23d6: Status 404 returned error can't find the container with id 1984cb08147761de40eee4fdcf3bcfa712e8bf8a7d1292247473d243de6b23d6 Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.161717 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d675855ff-hqhh5"] Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.894369 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" event={"ID":"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6","Type":"ContainerStarted","Data":"4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc"} Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.894836 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" event={"ID":"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6","Type":"ContainerStarted","Data":"1168add8bc3a930374818657e4fce87c6e87b9dda58d6de4722a0eb2c0cb8b9d"} Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.894878 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.896511 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" event={"ID":"32c36f41-337f-4a4d-b0ec-3e45091fbef2","Type":"ContainerStarted","Data":"aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216"} Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.896567 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" event={"ID":"32c36f41-337f-4a4d-b0ec-3e45091fbef2","Type":"ContainerStarted","Data":"1984cb08147761de40eee4fdcf3bcfa712e8bf8a7d1292247473d243de6b23d6"} Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.896818 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.899402 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vc5ws" event={"ID":"b0e828a1-a902-411f-8b55-f12b3a8c3e1c","Type":"ContainerStarted","Data":"eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee"} Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.902140 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fwhv" event={"ID":"c83fd1dd-24ef-45f8-a3ff-264d74862d94","Type":"ContainerStarted","Data":"2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05"} Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.904984 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.905022 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.916189 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" podStartSLOduration=2.916176112 podStartE2EDuration="2.916176112s" podCreationTimestamp="2026-02-20 06:45:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:45:38.915264665 +0000 UTC m=+295.686553643" watchObservedRunningTime="2026-02-20 06:45:38.916176112 +0000 UTC m=+295.687465090" Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.949713 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vc5ws" podStartSLOduration=2.4499199369999998 podStartE2EDuration="3.949700152s" podCreationTimestamp="2026-02-20 06:45:35 +0000 UTC" firstStartedPulling="2026-02-20 06:45:36.863512108 +0000 UTC m=+293.634801086" lastFinishedPulling="2026-02-20 06:45:38.363292322 +0000 UTC m=+295.134581301" observedRunningTime="2026-02-20 06:45:38.947302744 +0000 UTC m=+295.718591722" watchObservedRunningTime="2026-02-20 06:45:38.949700152 +0000 UTC m=+295.720989121" Feb 20 06:45:38 crc kubenswrapper[4492]: I0220 06:45:38.975458 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" podStartSLOduration=2.975447157 podStartE2EDuration="2.975447157s" podCreationTimestamp="2026-02-20 06:45:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:45:38.973382587 +0000 UTC m=+295.744671566" watchObservedRunningTime="2026-02-20 06:45:38.975447157 +0000 UTC m=+295.746736126" Feb 20 06:45:43 crc kubenswrapper[4492]: I0220 06:45:43.444808 4492 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.133535 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8fwhv" podStartSLOduration=7.660667097 podStartE2EDuration="9.133504293s" podCreationTimestamp="2026-02-20 06:45:35 +0000 UTC" firstStartedPulling="2026-02-20 06:45:36.873506901 +0000 UTC m=+293.644795879" lastFinishedPulling="2026-02-20 06:45:38.346344097 +0000 UTC m=+295.117633075" observedRunningTime="2026-02-20 06:45:38.995572421 +0000 UTC m=+295.766861399" watchObservedRunningTime="2026-02-20 06:45:44.133504293 +0000 UTC m=+300.904793271" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.134437 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d675855ff-hqhh5"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.134832 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" podUID="e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" containerName="controller-manager" containerID="cri-o://4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc" gracePeriod=30 Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.210735 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.211118 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" podUID="32c36f41-337f-4a4d-b0ec-3e45091fbef2" containerName="route-controller-manager" containerID="cri-o://aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216" gracePeriod=30 Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.551295 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.555735 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.606817 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-proxy-ca-bundles\") pod \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.606923 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-serving-cert\") pod \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.607648 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" (UID: "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.607983 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdzn4\" (UniqueName: \"kubernetes.io/projected/32c36f41-337f-4a4d-b0ec-3e45091fbef2-kube-api-access-sdzn4\") pod \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.608166 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltd6h\" (UniqueName: \"kubernetes.io/projected/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-kube-api-access-ltd6h\") pod \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.608201 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-config\") pod \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.608719 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-client-ca\") pod \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.608810 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-config" (OuterVolumeSpecName: "config") pod "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" (UID: "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609193 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-config\") pod \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609272 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-client-ca\") pod \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\" (UID: \"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609295 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c36f41-337f-4a4d-b0ec-3e45091fbef2-serving-cert\") pod \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\" (UID: \"32c36f41-337f-4a4d-b0ec-3e45091fbef2\") " Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609434 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-client-ca" (OuterVolumeSpecName: "client-ca") pod "32c36f41-337f-4a4d-b0ec-3e45091fbef2" (UID: "32c36f41-337f-4a4d-b0ec-3e45091fbef2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609885 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609907 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609918 4492 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.609964 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-config" (OuterVolumeSpecName: "config") pod "32c36f41-337f-4a4d-b0ec-3e45091fbef2" (UID: "32c36f41-337f-4a4d-b0ec-3e45091fbef2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.614563 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" (UID: "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.618734 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" (UID: "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.619071 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32c36f41-337f-4a4d-b0ec-3e45091fbef2-kube-api-access-sdzn4" (OuterVolumeSpecName: "kube-api-access-sdzn4") pod "32c36f41-337f-4a4d-b0ec-3e45091fbef2" (UID: "32c36f41-337f-4a4d-b0ec-3e45091fbef2"). InnerVolumeSpecName "kube-api-access-sdzn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.619248 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32c36f41-337f-4a4d-b0ec-3e45091fbef2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "32c36f41-337f-4a4d-b0ec-3e45091fbef2" (UID: "32c36f41-337f-4a4d-b0ec-3e45091fbef2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.620620 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-kube-api-access-ltd6h" (OuterVolumeSpecName: "kube-api-access-ltd6h") pod "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" (UID: "e591ae3b-9127-4cf1-9afa-a31d2f5a98a6"). InnerVolumeSpecName "kube-api-access-ltd6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.712187 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32c36f41-337f-4a4d-b0ec-3e45091fbef2-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.712220 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.712232 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32c36f41-337f-4a4d-b0ec-3e45091fbef2-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.712242 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.712254 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdzn4\" (UniqueName: \"kubernetes.io/projected/32c36f41-337f-4a4d-b0ec-3e45091fbef2-kube-api-access-sdzn4\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.712266 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltd6h\" (UniqueName: \"kubernetes.io/projected/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6-kube-api-access-ltd6h\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.798119 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-79748c675d-nwqvt"] Feb 20 06:45:44 crc kubenswrapper[4492]: E0220 06:45:44.798373 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" containerName="controller-manager" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.798387 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" containerName="controller-manager" Feb 20 06:45:44 crc kubenswrapper[4492]: E0220 06:45:44.798414 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32c36f41-337f-4a4d-b0ec-3e45091fbef2" containerName="route-controller-manager" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.798422 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="32c36f41-337f-4a4d-b0ec-3e45091fbef2" containerName="route-controller-manager" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.798588 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="32c36f41-337f-4a4d-b0ec-3e45091fbef2" containerName="route-controller-manager" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.798600 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" containerName="controller-manager" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.799119 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.814689 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79748c675d-nwqvt"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.900365 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.901216 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.914909 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.916755 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/584a19c8-b2e6-4175-b448-2098533040c1-serving-cert\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.916824 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgns2\" (UniqueName: \"kubernetes.io/projected/584a19c8-b2e6-4175-b448-2098533040c1-kube-api-access-zgns2\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.917045 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-proxy-ca-bundles\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.917173 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-client-ca\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.917285 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-config\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.936399 4492 generic.go:334] "Generic (PLEG): container finished" podID="32c36f41-337f-4a4d-b0ec-3e45091fbef2" containerID="aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216" exitCode=0 Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.936468 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" event={"ID":"32c36f41-337f-4a4d-b0ec-3e45091fbef2","Type":"ContainerDied","Data":"aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216"} Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.936535 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" event={"ID":"32c36f41-337f-4a4d-b0ec-3e45091fbef2","Type":"ContainerDied","Data":"1984cb08147761de40eee4fdcf3bcfa712e8bf8a7d1292247473d243de6b23d6"} Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.936547 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.936555 4492 scope.go:117] "RemoveContainer" containerID="aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.938460 4492 generic.go:334] "Generic (PLEG): container finished" podID="e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" containerID="4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc" exitCode=0 Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.938526 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" event={"ID":"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6","Type":"ContainerDied","Data":"4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc"} Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.938555 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" event={"ID":"e591ae3b-9127-4cf1-9afa-a31d2f5a98a6","Type":"ContainerDied","Data":"1168add8bc3a930374818657e4fce87c6e87b9dda58d6de4722a0eb2c0cb8b9d"} Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.938559 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d675855ff-hqhh5" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.960985 4492 scope.go:117] "RemoveContainer" containerID="aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216" Feb 20 06:45:44 crc kubenswrapper[4492]: E0220 06:45:44.961649 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216\": container with ID starting with aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216 not found: ID does not exist" containerID="aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.961680 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216"} err="failed to get container status \"aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216\": rpc error: code = NotFound desc = could not find container \"aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216\": container with ID starting with aaba3587a2388154f5c9834c530768e617c2e97dd4e9b4eccfa42dd0ec247216 not found: ID does not exist" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.961702 4492 scope.go:117] "RemoveContainer" containerID="4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.976522 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d675855ff-hqhh5"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.977868 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6d675855ff-hqhh5"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.980112 4492 scope.go:117] "RemoveContainer" containerID="4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc" Feb 20 06:45:44 crc kubenswrapper[4492]: E0220 06:45:44.980586 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc\": container with ID starting with 4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc not found: ID does not exist" containerID="4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.980615 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc"} err="failed to get container status \"4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc\": rpc error: code = NotFound desc = could not find container \"4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc\": container with ID starting with 4dbc55b92bb6f38b6b5ad2aac0373320ca702743a7da375381221dc3298703cc not found: ID does not exist" Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.985162 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls"] Feb 20 06:45:44 crc kubenswrapper[4492]: I0220 06:45:44.991418 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c6c76fcf7-6b2ls"] Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018535 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-config\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018577 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdlbl\" (UniqueName: \"kubernetes.io/projected/3ba3326a-250e-4d37-9fc4-3dc848f310c5-kube-api-access-vdlbl\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018619 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgns2\" (UniqueName: \"kubernetes.io/projected/584a19c8-b2e6-4175-b448-2098533040c1-kube-api-access-zgns2\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018643 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba3326a-250e-4d37-9fc4-3dc848f310c5-serving-cert\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018676 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-proxy-ca-bundles\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018705 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-client-ca\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018738 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-config\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018762 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-client-ca\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.018784 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/584a19c8-b2e6-4175-b448-2098533040c1-serving-cert\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.020724 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-proxy-ca-bundles\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.020989 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-client-ca\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.021203 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-config\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.024365 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/584a19c8-b2e6-4175-b448-2098533040c1-serving-cert\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.034679 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgns2\" (UniqueName: \"kubernetes.io/projected/584a19c8-b2e6-4175-b448-2098533040c1-kube-api-access-zgns2\") pod \"controller-manager-79748c675d-nwqvt\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.118163 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.119876 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba3326a-250e-4d37-9fc4-3dc848f310c5-serving-cert\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.120068 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-client-ca\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.120106 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-config\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.120145 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdlbl\" (UniqueName: \"kubernetes.io/projected/3ba3326a-250e-4d37-9fc4-3dc848f310c5-kube-api-access-vdlbl\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.121441 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-client-ca\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.122190 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-config\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.124067 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba3326a-250e-4d37-9fc4-3dc848f310c5-serving-cert\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.136839 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdlbl\" (UniqueName: \"kubernetes.io/projected/3ba3326a-250e-4d37-9fc4-3dc848f310c5-kube-api-access-vdlbl\") pod \"route-controller-manager-54769cd8f6-tr8r7\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.214999 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:45 crc kubenswrapper[4492]: I0220 06:45:45.333424 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-79748c675d-nwqvt"] Feb 20 06:45:45 crc kubenswrapper[4492]: W0220 06:45:45.343392 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod584a19c8_b2e6_4175_b448_2098533040c1.slice/crio-2ce3ef677424917be7b0c847c92ff5542bf458810a38eefc0d91122027b0102f WatchSource:0}: Error finding container 2ce3ef677424917be7b0c847c92ff5542bf458810a38eefc0d91122027b0102f: Status 404 returned error can't find the container with id 2ce3ef677424917be7b0c847c92ff5542bf458810a38eefc0d91122027b0102f Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.490950 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7"] Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.568153 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32c36f41-337f-4a4d-b0ec-3e45091fbef2" path="/var/lib/kubelet/pods/32c36f41-337f-4a4d-b0ec-3e45091fbef2/volumes" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.569136 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e591ae3b-9127-4cf1-9afa-a31d2f5a98a6" path="/var/lib/kubelet/pods/e591ae3b-9127-4cf1-9afa-a31d2f5a98a6/volumes" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.954720 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" event={"ID":"3ba3326a-250e-4d37-9fc4-3dc848f310c5","Type":"ContainerStarted","Data":"65768529a396db4efb4a85d6ea7e095728f2c662a1cb47a5c1ef4fba60ef1e86"} Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.955065 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.955081 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" event={"ID":"3ba3326a-250e-4d37-9fc4-3dc848f310c5","Type":"ContainerStarted","Data":"2a0285a6b714795c3042d0236f314a9f0585f12aae9ba22070e955338bb4c6d8"} Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.957715 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" event={"ID":"584a19c8-b2e6-4175-b448-2098533040c1","Type":"ContainerStarted","Data":"54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749"} Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.957743 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" event={"ID":"584a19c8-b2e6-4175-b448-2098533040c1","Type":"ContainerStarted","Data":"2ce3ef677424917be7b0c847c92ff5542bf458810a38eefc0d91122027b0102f"} Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.958058 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.965267 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.973378 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" podStartSLOduration=1.973362844 podStartE2EDuration="1.973362844s" podCreationTimestamp="2026-02-20 06:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:45:45.970797339 +0000 UTC m=+302.742086327" watchObservedRunningTime="2026-02-20 06:45:45.973362844 +0000 UTC m=+302.744651822" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.993315 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" podStartSLOduration=1.993274234 podStartE2EDuration="1.993274234s" podCreationTimestamp="2026-02-20 06:45:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:45:45.987996849 +0000 UTC m=+302.759285826" watchObservedRunningTime="2026-02-20 06:45:45.993274234 +0000 UTC m=+302.764563211" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:45.994109 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:46.062898 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:46.072897 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:46.114566 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:46.260640 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:46.260694 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:46 crc kubenswrapper[4492]: I0220 06:45:46.290349 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:47 crc kubenswrapper[4492]: I0220 06:45:47.003597 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 06:45:47 crc kubenswrapper[4492]: I0220 06:45:47.003959 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 06:45:56 crc kubenswrapper[4492]: I0220 06:45:56.829784 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7"] Feb 20 06:45:56 crc kubenswrapper[4492]: I0220 06:45:56.830626 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" podUID="3ba3326a-250e-4d37-9fc4-3dc848f310c5" containerName="route-controller-manager" containerID="cri-o://65768529a396db4efb4a85d6ea7e095728f2c662a1cb47a5c1ef4fba60ef1e86" gracePeriod=30 Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.022888 4492 generic.go:334] "Generic (PLEG): container finished" podID="3ba3326a-250e-4d37-9fc4-3dc848f310c5" containerID="65768529a396db4efb4a85d6ea7e095728f2c662a1cb47a5c1ef4fba60ef1e86" exitCode=0 Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.022976 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" event={"ID":"3ba3326a-250e-4d37-9fc4-3dc848f310c5","Type":"ContainerDied","Data":"65768529a396db4efb4a85d6ea7e095728f2c662a1cb47a5c1ef4fba60ef1e86"} Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.254893 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.298800 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba3326a-250e-4d37-9fc4-3dc848f310c5-serving-cert\") pod \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.298858 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-client-ca\") pod \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.298970 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdlbl\" (UniqueName: \"kubernetes.io/projected/3ba3326a-250e-4d37-9fc4-3dc848f310c5-kube-api-access-vdlbl\") pod \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.299002 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-config\") pod \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\" (UID: \"3ba3326a-250e-4d37-9fc4-3dc848f310c5\") " Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.299776 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-config" (OuterVolumeSpecName: "config") pod "3ba3326a-250e-4d37-9fc4-3dc848f310c5" (UID: "3ba3326a-250e-4d37-9fc4-3dc848f310c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.300001 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-client-ca" (OuterVolumeSpecName: "client-ca") pod "3ba3326a-250e-4d37-9fc4-3dc848f310c5" (UID: "3ba3326a-250e-4d37-9fc4-3dc848f310c5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.310797 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ba3326a-250e-4d37-9fc4-3dc848f310c5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3ba3326a-250e-4d37-9fc4-3dc848f310c5" (UID: "3ba3326a-250e-4d37-9fc4-3dc848f310c5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.311823 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ba3326a-250e-4d37-9fc4-3dc848f310c5-kube-api-access-vdlbl" (OuterVolumeSpecName: "kube-api-access-vdlbl") pod "3ba3326a-250e-4d37-9fc4-3dc848f310c5" (UID: "3ba3326a-250e-4d37-9fc4-3dc848f310c5"). InnerVolumeSpecName "kube-api-access-vdlbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.400798 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdlbl\" (UniqueName: \"kubernetes.io/projected/3ba3326a-250e-4d37-9fc4-3dc848f310c5-kube-api-access-vdlbl\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.400840 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.400855 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba3326a-250e-4d37-9fc4-3dc848f310c5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:57 crc kubenswrapper[4492]: I0220 06:45:57.400865 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3ba3326a-250e-4d37-9fc4-3dc848f310c5-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.029209 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" event={"ID":"3ba3326a-250e-4d37-9fc4-3dc848f310c5","Type":"ContainerDied","Data":"2a0285a6b714795c3042d0236f314a9f0585f12aae9ba22070e955338bb4c6d8"} Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.029624 4492 scope.go:117] "RemoveContainer" containerID="65768529a396db4efb4a85d6ea7e095728f2c662a1cb47a5c1ef4fba60ef1e86" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.029260 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.042753 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7"] Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.057863 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54769cd8f6-tr8r7"] Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.608710 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf"] Feb 20 06:45:58 crc kubenswrapper[4492]: E0220 06:45:58.609063 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba3326a-250e-4d37-9fc4-3dc848f310c5" containerName="route-controller-manager" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.609088 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba3326a-250e-4d37-9fc4-3dc848f310c5" containerName="route-controller-manager" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.609239 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ba3326a-250e-4d37-9fc4-3dc848f310c5" containerName="route-controller-manager" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.609922 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.611983 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.612344 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.612615 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.612702 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.612897 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.613175 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.623985 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf"] Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.716459 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0109fd3-e179-4b71-a3f3-5d8f17b35931-serving-cert\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.716567 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0109fd3-e179-4b71-a3f3-5d8f17b35931-config\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.716621 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgvt6\" (UniqueName: \"kubernetes.io/projected/e0109fd3-e179-4b71-a3f3-5d8f17b35931-kube-api-access-jgvt6\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.716664 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0109fd3-e179-4b71-a3f3-5d8f17b35931-client-ca\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.818922 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0109fd3-e179-4b71-a3f3-5d8f17b35931-serving-cert\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.818985 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0109fd3-e179-4b71-a3f3-5d8f17b35931-config\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.819028 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgvt6\" (UniqueName: \"kubernetes.io/projected/e0109fd3-e179-4b71-a3f3-5d8f17b35931-kube-api-access-jgvt6\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.819065 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0109fd3-e179-4b71-a3f3-5d8f17b35931-client-ca\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.820602 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0109fd3-e179-4b71-a3f3-5d8f17b35931-client-ca\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.820776 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0109fd3-e179-4b71-a3f3-5d8f17b35931-config\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.823957 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0109fd3-e179-4b71-a3f3-5d8f17b35931-serving-cert\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.833198 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgvt6\" (UniqueName: \"kubernetes.io/projected/e0109fd3-e179-4b71-a3f3-5d8f17b35931-kube-api-access-jgvt6\") pod \"route-controller-manager-6b9bdfbfb-xtgxf\" (UID: \"e0109fd3-e179-4b71-a3f3-5d8f17b35931\") " pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:58 crc kubenswrapper[4492]: I0220 06:45:58.924108 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.114233 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qp2hh"] Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.130838 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.131005 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qp2hh"] Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.234528 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-registry-tls\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.234829 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.234881 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-registry-certificates\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.234911 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-trusted-ca\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.234938 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.234968 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-bound-sa-token\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.234996 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rndjk\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-kube-api-access-rndjk\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.235076 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.253874 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.336518 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-registry-tls\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.336604 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.336667 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-registry-certificates\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.336719 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-trusted-ca\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.336773 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.336838 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-bound-sa-token\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.336873 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rndjk\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-kube-api-access-rndjk\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.337353 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.337997 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-registry-certificates\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.338082 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-trusted-ca\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.342355 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-registry-tls\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.342921 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.352093 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rndjk\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-kube-api-access-rndjk\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.352637 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b70f43e-a732-4f3d-b149-72b6c0a8c53c-bound-sa-token\") pod \"image-registry-66df7c8f76-qp2hh\" (UID: \"2b70f43e-a732-4f3d-b149-72b6c0a8c53c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.379741 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf"] Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.460220 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.568974 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ba3326a-250e-4d37-9fc4-3dc848f310c5" path="/var/lib/kubelet/pods/3ba3326a-250e-4d37-9fc4-3dc848f310c5/volumes" Feb 20 06:45:59 crc kubenswrapper[4492]: I0220 06:45:59.890266 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qp2hh"] Feb 20 06:45:59 crc kubenswrapper[4492]: W0220 06:45:59.903029 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b70f43e_a732_4f3d_b149_72b6c0a8c53c.slice/crio-79b246cd7745c04152487c05a2460e4d4538ddeb8c63bdd2b0dce8b46ae8f723 WatchSource:0}: Error finding container 79b246cd7745c04152487c05a2460e4d4538ddeb8c63bdd2b0dce8b46ae8f723: Status 404 returned error can't find the container with id 79b246cd7745c04152487c05a2460e4d4538ddeb8c63bdd2b0dce8b46ae8f723 Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.047815 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" event={"ID":"2b70f43e-a732-4f3d-b149-72b6c0a8c53c","Type":"ContainerStarted","Data":"142d08712e968df487c82ff9891383156915d1d169cafd1ca58fd4e0e3ac3ee6"} Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.048265 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" event={"ID":"2b70f43e-a732-4f3d-b149-72b6c0a8c53c","Type":"ContainerStarted","Data":"79b246cd7745c04152487c05a2460e4d4538ddeb8c63bdd2b0dce8b46ae8f723"} Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.048331 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.056383 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" event={"ID":"e0109fd3-e179-4b71-a3f3-5d8f17b35931","Type":"ContainerStarted","Data":"ed95c132b0e04f79d06ecd5130641d79d26b035a50c523286b18d0541fcfd20c"} Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.056418 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" event={"ID":"e0109fd3-e179-4b71-a3f3-5d8f17b35931","Type":"ContainerStarted","Data":"e4c439b570a1c5a80d98b7074ab505c06bda379eee9e5d4c197bc87f0e91f429"} Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.056706 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.062946 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" Feb 20 06:46:00 crc kubenswrapper[4492]: I0220 06:46:00.075662 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" podStartSLOduration=1.075650724 podStartE2EDuration="1.075650724s" podCreationTimestamp="2026-02-20 06:45:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:46:00.073941162 +0000 UTC m=+316.845230140" watchObservedRunningTime="2026-02-20 06:46:00.075650724 +0000 UTC m=+316.846939701" Feb 20 06:46:09 crc kubenswrapper[4492]: I0220 06:46:09.311432 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:46:09 crc kubenswrapper[4492]: I0220 06:46:09.312023 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.340041 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b9bdfbfb-xtgxf" podStartSLOduration=20.340026226 podStartE2EDuration="20.340026226s" podCreationTimestamp="2026-02-20 06:45:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:46:00.10838205 +0000 UTC m=+316.879671028" watchObservedRunningTime="2026-02-20 06:46:16.340026226 +0000 UTC m=+333.111315215" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.341447 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79748c675d-nwqvt"] Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.341636 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" podUID="584a19c8-b2e6-4175-b448-2098533040c1" containerName="controller-manager" containerID="cri-o://54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749" gracePeriod=30 Feb 20 06:46:16 crc kubenswrapper[4492]: E0220 06:46:16.380822 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod584a19c8_b2e6_4175_b448_2098533040c1.slice/crio-conmon-54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749.scope\": RecentStats: unable to find data in memory cache]" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.806616 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.848336 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-config\") pod \"584a19c8-b2e6-4175-b448-2098533040c1\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.848394 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-proxy-ca-bundles\") pod \"584a19c8-b2e6-4175-b448-2098533040c1\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.848437 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-client-ca\") pod \"584a19c8-b2e6-4175-b448-2098533040c1\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.848492 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgns2\" (UniqueName: \"kubernetes.io/projected/584a19c8-b2e6-4175-b448-2098533040c1-kube-api-access-zgns2\") pod \"584a19c8-b2e6-4175-b448-2098533040c1\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.848515 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/584a19c8-b2e6-4175-b448-2098533040c1-serving-cert\") pod \"584a19c8-b2e6-4175-b448-2098533040c1\" (UID: \"584a19c8-b2e6-4175-b448-2098533040c1\") " Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.849090 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-client-ca" (OuterVolumeSpecName: "client-ca") pod "584a19c8-b2e6-4175-b448-2098533040c1" (UID: "584a19c8-b2e6-4175-b448-2098533040c1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.849141 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-config" (OuterVolumeSpecName: "config") pod "584a19c8-b2e6-4175-b448-2098533040c1" (UID: "584a19c8-b2e6-4175-b448-2098533040c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.849366 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "584a19c8-b2e6-4175-b448-2098533040c1" (UID: "584a19c8-b2e6-4175-b448-2098533040c1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.854343 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/584a19c8-b2e6-4175-b448-2098533040c1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "584a19c8-b2e6-4175-b448-2098533040c1" (UID: "584a19c8-b2e6-4175-b448-2098533040c1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.854423 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584a19c8-b2e6-4175-b448-2098533040c1-kube-api-access-zgns2" (OuterVolumeSpecName: "kube-api-access-zgns2") pod "584a19c8-b2e6-4175-b448-2098533040c1" (UID: "584a19c8-b2e6-4175-b448-2098533040c1"). InnerVolumeSpecName "kube-api-access-zgns2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.949263 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.949288 4492 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.949299 4492 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/584a19c8-b2e6-4175-b448-2098533040c1-client-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.949307 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgns2\" (UniqueName: \"kubernetes.io/projected/584a19c8-b2e6-4175-b448-2098533040c1-kube-api-access-zgns2\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:16 crc kubenswrapper[4492]: I0220 06:46:16.949316 4492 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/584a19c8-b2e6-4175-b448-2098533040c1-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.128020 4492 generic.go:334] "Generic (PLEG): container finished" podID="584a19c8-b2e6-4175-b448-2098533040c1" containerID="54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749" exitCode=0 Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.128059 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" event={"ID":"584a19c8-b2e6-4175-b448-2098533040c1","Type":"ContainerDied","Data":"54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749"} Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.128073 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.128259 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-79748c675d-nwqvt" event={"ID":"584a19c8-b2e6-4175-b448-2098533040c1","Type":"ContainerDied","Data":"2ce3ef677424917be7b0c847c92ff5542bf458810a38eefc0d91122027b0102f"} Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.128283 4492 scope.go:117] "RemoveContainer" containerID="54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.142034 4492 scope.go:117] "RemoveContainer" containerID="54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749" Feb 20 06:46:17 crc kubenswrapper[4492]: E0220 06:46:17.142430 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749\": container with ID starting with 54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749 not found: ID does not exist" containerID="54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.142456 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749"} err="failed to get container status \"54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749\": rpc error: code = NotFound desc = could not find container \"54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749\": container with ID starting with 54059dbe5ed24eb2a6c8fad54b50af3adf824ec20a8dddcfc69667137aeb3749 not found: ID does not exist" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.146784 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-79748c675d-nwqvt"] Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.150669 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-79748c675d-nwqvt"] Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.562353 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="584a19c8-b2e6-4175-b448-2098533040c1" path="/var/lib/kubelet/pods/584a19c8-b2e6-4175-b448-2098533040c1/volumes" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.617102 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7976878f64-228nv"] Feb 20 06:46:17 crc kubenswrapper[4492]: E0220 06:46:17.617301 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584a19c8-b2e6-4175-b448-2098533040c1" containerName="controller-manager" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.617317 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="584a19c8-b2e6-4175-b448-2098533040c1" containerName="controller-manager" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.617412 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="584a19c8-b2e6-4175-b448-2098533040c1" containerName="controller-manager" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.617761 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.619724 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.620046 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.620307 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.620428 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.620640 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.622248 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.628879 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.630782 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7976878f64-228nv"] Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.658104 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-client-ca\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.658151 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb5wf\" (UniqueName: \"kubernetes.io/projected/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-kube-api-access-vb5wf\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.658216 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-serving-cert\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.658238 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-config\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.658358 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-proxy-ca-bundles\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.759034 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-serving-cert\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.759066 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-config\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.759121 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-proxy-ca-bundles\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.759142 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-client-ca\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.759165 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb5wf\" (UniqueName: \"kubernetes.io/projected/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-kube-api-access-vb5wf\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.760084 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-client-ca\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.760280 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-proxy-ca-bundles\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.760348 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-config\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.763541 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-serving-cert\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.771807 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb5wf\" (UniqueName: \"kubernetes.io/projected/4efbe1d6-7bc2-41c6-b738-018e4ce376cb-kube-api-access-vb5wf\") pod \"controller-manager-7976878f64-228nv\" (UID: \"4efbe1d6-7bc2-41c6-b738-018e4ce376cb\") " pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:17 crc kubenswrapper[4492]: I0220 06:46:17.929300 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:18 crc kubenswrapper[4492]: I0220 06:46:18.270685 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7976878f64-228nv"] Feb 20 06:46:19 crc kubenswrapper[4492]: I0220 06:46:19.139574 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7976878f64-228nv" event={"ID":"4efbe1d6-7bc2-41c6-b738-018e4ce376cb","Type":"ContainerStarted","Data":"f0eb1cbce0a3aaccb98e94cee42f468de02ccf3f2ebd277e2eb26985920bb9cc"} Feb 20 06:46:19 crc kubenswrapper[4492]: I0220 06:46:19.139622 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7976878f64-228nv" event={"ID":"4efbe1d6-7bc2-41c6-b738-018e4ce376cb","Type":"ContainerStarted","Data":"c221a5d7ed01b7460b595efcf2eaf7716f27837116c832322005d95a20b1b39a"} Feb 20 06:46:19 crc kubenswrapper[4492]: I0220 06:46:19.139733 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:19 crc kubenswrapper[4492]: I0220 06:46:19.143003 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7976878f64-228nv" Feb 20 06:46:19 crc kubenswrapper[4492]: I0220 06:46:19.152720 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7976878f64-228nv" podStartSLOduration=3.15270941 podStartE2EDuration="3.15270941s" podCreationTimestamp="2026-02-20 06:46:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:46:19.152070036 +0000 UTC m=+335.923359013" watchObservedRunningTime="2026-02-20 06:46:19.15270941 +0000 UTC m=+335.923998388" Feb 20 06:46:19 crc kubenswrapper[4492]: I0220 06:46:19.466446 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-qp2hh" Feb 20 06:46:19 crc kubenswrapper[4492]: I0220 06:46:19.505785 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t9jrw"] Feb 20 06:46:39 crc kubenswrapper[4492]: I0220 06:46:39.311180 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:46:39 crc kubenswrapper[4492]: I0220 06:46:39.311770 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.532942 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" podUID="bd14138a-551f-4bbf-9934-b7555cb1d5e6" containerName="registry" containerID="cri-o://1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a" gracePeriod=30 Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.943075 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970554 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-bound-sa-token\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970600 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5n2w\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-kube-api-access-b5n2w\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970659 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-tls\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970784 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970822 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-trusted-ca\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970881 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd14138a-551f-4bbf-9934-b7555cb1d5e6-ca-trust-extracted\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970932 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-certificates\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.970979 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd14138a-551f-4bbf-9934-b7555cb1d5e6-installation-pull-secrets\") pod \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\" (UID: \"bd14138a-551f-4bbf-9934-b7555cb1d5e6\") " Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.971563 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.971605 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.976725 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd14138a-551f-4bbf-9934-b7555cb1d5e6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.978205 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-kube-api-access-b5n2w" (OuterVolumeSpecName: "kube-api-access-b5n2w") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "kube-api-access-b5n2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.978433 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.978495 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.981209 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:46:44 crc kubenswrapper[4492]: I0220 06:46:44.984336 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd14138a-551f-4bbf-9934-b7555cb1d5e6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "bd14138a-551f-4bbf-9934-b7555cb1d5e6" (UID: "bd14138a-551f-4bbf-9934-b7555cb1d5e6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.072179 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5n2w\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-kube-api-access-b5n2w\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.072218 4492 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.072234 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.072245 4492 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/bd14138a-551f-4bbf-9934-b7555cb1d5e6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.072256 4492 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/bd14138a-551f-4bbf-9934-b7555cb1d5e6-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.072269 4492 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/bd14138a-551f-4bbf-9934-b7555cb1d5e6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.072284 4492 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bd14138a-551f-4bbf-9934-b7555cb1d5e6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.280838 4492 generic.go:334] "Generic (PLEG): container finished" podID="bd14138a-551f-4bbf-9934-b7555cb1d5e6" containerID="1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a" exitCode=0 Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.280908 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" event={"ID":"bd14138a-551f-4bbf-9934-b7555cb1d5e6","Type":"ContainerDied","Data":"1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a"} Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.281196 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" event={"ID":"bd14138a-551f-4bbf-9934-b7555cb1d5e6","Type":"ContainerDied","Data":"873768df4a6f995f258639052b26a08756246c7f435cbd1afc76f152e74559aa"} Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.281222 4492 scope.go:117] "RemoveContainer" containerID="1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.280967 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-t9jrw" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.296621 4492 scope.go:117] "RemoveContainer" containerID="1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a" Feb 20 06:46:45 crc kubenswrapper[4492]: E0220 06:46:45.297199 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a\": container with ID starting with 1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a not found: ID does not exist" containerID="1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.297253 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a"} err="failed to get container status \"1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a\": rpc error: code = NotFound desc = could not find container \"1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a\": container with ID starting with 1013696873cf8d9535939f0e3d8c2e091aea4b6cdaa1144c02a15df56fd68c4a not found: ID does not exist" Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.309827 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t9jrw"] Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.316882 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-t9jrw"] Feb 20 06:46:45 crc kubenswrapper[4492]: I0220 06:46:45.563436 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd14138a-551f-4bbf-9934-b7555cb1d5e6" path="/var/lib/kubelet/pods/bd14138a-551f-4bbf-9934-b7555cb1d5e6/volumes" Feb 20 06:47:09 crc kubenswrapper[4492]: I0220 06:47:09.311939 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:47:09 crc kubenswrapper[4492]: I0220 06:47:09.312679 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:47:09 crc kubenswrapper[4492]: I0220 06:47:09.312745 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:47:09 crc kubenswrapper[4492]: I0220 06:47:09.313365 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f0bbbbef7d028667e57fa52189868db9e398ca3ba68113428d3adebbf972641"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 06:47:09 crc kubenswrapper[4492]: I0220 06:47:09.313438 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://8f0bbbbef7d028667e57fa52189868db9e398ca3ba68113428d3adebbf972641" gracePeriod=600 Feb 20 06:47:10 crc kubenswrapper[4492]: I0220 06:47:10.437741 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="8f0bbbbef7d028667e57fa52189868db9e398ca3ba68113428d3adebbf972641" exitCode=0 Feb 20 06:47:10 crc kubenswrapper[4492]: I0220 06:47:10.437839 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"8f0bbbbef7d028667e57fa52189868db9e398ca3ba68113428d3adebbf972641"} Feb 20 06:47:10 crc kubenswrapper[4492]: I0220 06:47:10.438533 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"855da3473a6d9146a28853331dfc686fe2c8735ee6afb03897c4083de78fb717"} Feb 20 06:47:10 crc kubenswrapper[4492]: I0220 06:47:10.438569 4492 scope.go:117] "RemoveContainer" containerID="81e2ac784e8a783bb8a1dbee4ccd916e249cb8572715afbf63e345398d063fbf" Feb 20 06:49:09 crc kubenswrapper[4492]: I0220 06:49:09.311847 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:49:09 crc kubenswrapper[4492]: I0220 06:49:09.312514 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.953975 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-68zkr"] Feb 20 06:49:13 crc kubenswrapper[4492]: E0220 06:49:13.956036 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd14138a-551f-4bbf-9934-b7555cb1d5e6" containerName="registry" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.956063 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd14138a-551f-4bbf-9934-b7555cb1d5e6" containerName="registry" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.956211 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd14138a-551f-4bbf-9934-b7555cb1d5e6" containerName="registry" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.956656 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.958922 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.962461 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-fgbwd"] Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.963135 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.963373 4492 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jpccn" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.967945 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-fgbwd" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.970653 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-68zkr"] Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.974096 4492 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-rkqpz" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.978167 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-l7wk4"] Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.978853 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.981519 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-fgbwd"] Feb 20 06:49:13 crc kubenswrapper[4492]: I0220 06:49:13.989054 4492 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-8tkmp" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.006072 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-l7wk4"] Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.106260 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9l6d\" (UniqueName: \"kubernetes.io/projected/b43a3a53-5167-477a-9531-ce42d03c5494-kube-api-access-t9l6d\") pod \"cert-manager-webhook-687f57d79b-l7wk4\" (UID: \"b43a3a53-5167-477a-9531-ce42d03c5494\") " pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.106322 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc979\" (UniqueName: \"kubernetes.io/projected/3e8ca167-0346-42b9-8e8e-48b1c85b3f5a-kube-api-access-gc979\") pod \"cert-manager-cainjector-cf98fcc89-68zkr\" (UID: \"3e8ca167-0346-42b9-8e8e-48b1c85b3f5a\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.106364 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxc6t\" (UniqueName: \"kubernetes.io/projected/63b72bed-a101-41fc-8402-f12eb5a6420b-kube-api-access-xxc6t\") pod \"cert-manager-858654f9db-fgbwd\" (UID: \"63b72bed-a101-41fc-8402-f12eb5a6420b\") " pod="cert-manager/cert-manager-858654f9db-fgbwd" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.207629 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9l6d\" (UniqueName: \"kubernetes.io/projected/b43a3a53-5167-477a-9531-ce42d03c5494-kube-api-access-t9l6d\") pod \"cert-manager-webhook-687f57d79b-l7wk4\" (UID: \"b43a3a53-5167-477a-9531-ce42d03c5494\") " pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.207710 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc979\" (UniqueName: \"kubernetes.io/projected/3e8ca167-0346-42b9-8e8e-48b1c85b3f5a-kube-api-access-gc979\") pod \"cert-manager-cainjector-cf98fcc89-68zkr\" (UID: \"3e8ca167-0346-42b9-8e8e-48b1c85b3f5a\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.207818 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxc6t\" (UniqueName: \"kubernetes.io/projected/63b72bed-a101-41fc-8402-f12eb5a6420b-kube-api-access-xxc6t\") pod \"cert-manager-858654f9db-fgbwd\" (UID: \"63b72bed-a101-41fc-8402-f12eb5a6420b\") " pod="cert-manager/cert-manager-858654f9db-fgbwd" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.228883 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc979\" (UniqueName: \"kubernetes.io/projected/3e8ca167-0346-42b9-8e8e-48b1c85b3f5a-kube-api-access-gc979\") pod \"cert-manager-cainjector-cf98fcc89-68zkr\" (UID: \"3e8ca167-0346-42b9-8e8e-48b1c85b3f5a\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.229756 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9l6d\" (UniqueName: \"kubernetes.io/projected/b43a3a53-5167-477a-9531-ce42d03c5494-kube-api-access-t9l6d\") pod \"cert-manager-webhook-687f57d79b-l7wk4\" (UID: \"b43a3a53-5167-477a-9531-ce42d03c5494\") " pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.229883 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxc6t\" (UniqueName: \"kubernetes.io/projected/63b72bed-a101-41fc-8402-f12eb5a6420b-kube-api-access-xxc6t\") pod \"cert-manager-858654f9db-fgbwd\" (UID: \"63b72bed-a101-41fc-8402-f12eb5a6420b\") " pod="cert-manager/cert-manager-858654f9db-fgbwd" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.274081 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.291118 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-fgbwd" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.317627 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.482556 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-68zkr"] Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.498331 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.522643 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-fgbwd"] Feb 20 06:49:14 crc kubenswrapper[4492]: W0220 06:49:14.537588 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63b72bed_a101_41fc_8402_f12eb5a6420b.slice/crio-3cc7c2143cd15715c987f17f3951810bb88ff046e8d60932a177944fa22d210b WatchSource:0}: Error finding container 3cc7c2143cd15715c987f17f3951810bb88ff046e8d60932a177944fa22d210b: Status 404 returned error can't find the container with id 3cc7c2143cd15715c987f17f3951810bb88ff046e8d60932a177944fa22d210b Feb 20 06:49:14 crc kubenswrapper[4492]: I0220 06:49:14.553019 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-l7wk4"] Feb 20 06:49:14 crc kubenswrapper[4492]: W0220 06:49:14.555458 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb43a3a53_5167_477a_9531_ce42d03c5494.slice/crio-f8c61d75535a8cbdf76a2e3f29a3ee405a36e25c359b3d4e54f35c6a01bc624d WatchSource:0}: Error finding container f8c61d75535a8cbdf76a2e3f29a3ee405a36e25c359b3d4e54f35c6a01bc624d: Status 404 returned error can't find the container with id f8c61d75535a8cbdf76a2e3f29a3ee405a36e25c359b3d4e54f35c6a01bc624d Feb 20 06:49:15 crc kubenswrapper[4492]: I0220 06:49:15.190580 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" event={"ID":"3e8ca167-0346-42b9-8e8e-48b1c85b3f5a","Type":"ContainerStarted","Data":"23d2d6b6db775dab64fed79d0450fa8396f1bcbdfaa6ff0ffec334a906527cad"} Feb 20 06:49:15 crc kubenswrapper[4492]: I0220 06:49:15.192335 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" event={"ID":"b43a3a53-5167-477a-9531-ce42d03c5494","Type":"ContainerStarted","Data":"f8c61d75535a8cbdf76a2e3f29a3ee405a36e25c359b3d4e54f35c6a01bc624d"} Feb 20 06:49:15 crc kubenswrapper[4492]: I0220 06:49:15.194087 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-fgbwd" event={"ID":"63b72bed-a101-41fc-8402-f12eb5a6420b","Type":"ContainerStarted","Data":"3cc7c2143cd15715c987f17f3951810bb88ff046e8d60932a177944fa22d210b"} Feb 20 06:49:18 crc kubenswrapper[4492]: I0220 06:49:18.214246 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-fgbwd" event={"ID":"63b72bed-a101-41fc-8402-f12eb5a6420b","Type":"ContainerStarted","Data":"6e46e4209701e066eb7f230be513bd3a3de19d437ddd8b9bbfdefcdfc5297cc8"} Feb 20 06:49:18 crc kubenswrapper[4492]: I0220 06:49:18.218556 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" event={"ID":"3e8ca167-0346-42b9-8e8e-48b1c85b3f5a","Type":"ContainerStarted","Data":"22cfa5736904e360207981f772c4833ad04f2dacc78b079a596110f041c495a4"} Feb 20 06:49:18 crc kubenswrapper[4492]: I0220 06:49:18.220551 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" event={"ID":"b43a3a53-5167-477a-9531-ce42d03c5494","Type":"ContainerStarted","Data":"447b84cd102bb592cbc24b5c43d14855104925bbf1a3fc0f4f1dc23e29a464be"} Feb 20 06:49:18 crc kubenswrapper[4492]: I0220 06:49:18.220666 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" Feb 20 06:49:18 crc kubenswrapper[4492]: I0220 06:49:18.235577 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-fgbwd" podStartSLOduration=1.9695866419999999 podStartE2EDuration="5.235561629s" podCreationTimestamp="2026-02-20 06:49:13 +0000 UTC" firstStartedPulling="2026-02-20 06:49:14.540739799 +0000 UTC m=+511.312028767" lastFinishedPulling="2026-02-20 06:49:17.806714776 +0000 UTC m=+514.578003754" observedRunningTime="2026-02-20 06:49:18.229066182 +0000 UTC m=+515.000355161" watchObservedRunningTime="2026-02-20 06:49:18.235561629 +0000 UTC m=+515.006850598" Feb 20 06:49:18 crc kubenswrapper[4492]: I0220 06:49:18.299262 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-68zkr" podStartSLOduration=2.029544106 podStartE2EDuration="5.299239943s" podCreationTimestamp="2026-02-20 06:49:13 +0000 UTC" firstStartedPulling="2026-02-20 06:49:14.498013711 +0000 UTC m=+511.269302689" lastFinishedPulling="2026-02-20 06:49:17.767709547 +0000 UTC m=+514.538998526" observedRunningTime="2026-02-20 06:49:18.256876167 +0000 UTC m=+515.028165144" watchObservedRunningTime="2026-02-20 06:49:18.299239943 +0000 UTC m=+515.070528920" Feb 20 06:49:18 crc kubenswrapper[4492]: I0220 06:49:18.306842 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" podStartSLOduration=2.104511638 podStartE2EDuration="5.306819652s" podCreationTimestamp="2026-02-20 06:49:13 +0000 UTC" firstStartedPulling="2026-02-20 06:49:14.558029043 +0000 UTC m=+511.329318021" lastFinishedPulling="2026-02-20 06:49:17.760337057 +0000 UTC m=+514.531626035" observedRunningTime="2026-02-20 06:49:18.292150886 +0000 UTC m=+515.063439874" watchObservedRunningTime="2026-02-20 06:49:18.306819652 +0000 UTC m=+515.078108630" Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.320005 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-l7wk4" Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.671425 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f68mj"] Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.671790 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-controller" containerID="cri-o://c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a" gracePeriod=30 Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.672169 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="sbdb" containerID="cri-o://5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0" gracePeriod=30 Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.672218 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="nbdb" containerID="cri-o://3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3" gracePeriod=30 Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.672253 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="northd" containerID="cri-o://8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a" gracePeriod=30 Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.672284 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c" gracePeriod=30 Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.672319 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-node" containerID="cri-o://5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622" gracePeriod=30 Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.672349 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-acl-logging" containerID="cri-o://9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25" gracePeriod=30 Feb 20 06:49:24 crc kubenswrapper[4492]: I0220 06:49:24.709238 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" containerID="cri-o://74a35c1868a572cbfcd7507555baf61f36d97d28af5ac21d481547bd4c44f779" gracePeriod=30 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.261682 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/2.log" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.262383 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/1.log" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.262430 4492 generic.go:334] "Generic (PLEG): container finished" podID="dcc11973-022a-47f9-b1e2-23e945352e7f" containerID="e63ed096fe97531f09118ad4101afeac3694e124f9060047cedcde53421e659a" exitCode=2 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.262517 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerDied","Data":"e63ed096fe97531f09118ad4101afeac3694e124f9060047cedcde53421e659a"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.262560 4492 scope.go:117] "RemoveContainer" containerID="d87ffc277bb282115504f5cbcf1d072dc053cbf2831416b74b77e6bc0dd285a8" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.263151 4492 scope.go:117] "RemoveContainer" containerID="e63ed096fe97531f09118ad4101afeac3694e124f9060047cedcde53421e659a" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.263431 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2rr9j_openshift-multus(dcc11973-022a-47f9-b1e2-23e945352e7f)\"" pod="openshift-multus/multus-2rr9j" podUID="dcc11973-022a-47f9-b1e2-23e945352e7f" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.264757 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovnkube-controller/3.log" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.266879 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovn-acl-logging/0.log" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267321 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovn-controller/0.log" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267674 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="74a35c1868a572cbfcd7507555baf61f36d97d28af5ac21d481547bd4c44f779" exitCode=0 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267696 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0" exitCode=0 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267703 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3" exitCode=0 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267709 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a" exitCode=0 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267715 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c" exitCode=0 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267721 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622" exitCode=0 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267727 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25" exitCode=143 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267904 4492 generic.go:334] "Generic (PLEG): container finished" podID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerID="c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a" exitCode=143 Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267923 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"74a35c1868a572cbfcd7507555baf61f36d97d28af5ac21d481547bd4c44f779"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267946 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267957 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267965 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267973 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267983 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267991 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.267999 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a"} Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.283355 4492 scope.go:117] "RemoveContainer" containerID="b496e7c0f791038edd742f3a76e3f3be69c7710b85503fcfda60cdaaa1be6238" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.340961 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovn-acl-logging/0.log" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.341274 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovn-controller/0.log" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.341561 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381372 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9pbsv"] Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381620 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="sbdb" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381642 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="sbdb" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381652 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381660 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381670 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="nbdb" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381675 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="nbdb" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381684 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-acl-logging" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381692 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-acl-logging" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381702 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381707 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381713 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381718 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381727 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="northd" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381732 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="northd" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381741 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-node" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381748 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-node" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381757 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kubecfg-setup" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381764 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kubecfg-setup" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381773 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381778 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381786 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381792 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.381798 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381803 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381886 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="nbdb" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381895 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381902 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-acl-logging" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381910 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381916 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="sbdb" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381923 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381930 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381936 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="northd" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381944 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381951 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-node" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381960 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovn-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.381966 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="kube-rbac-proxy-ovn-metrics" Feb 20 06:49:25 crc kubenswrapper[4492]: E0220 06:49:25.382053 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.382062 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" containerName="ovnkube-controller" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.383597 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459219 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-systemd\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459266 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-openvswitch\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459287 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-ovn-kubernetes\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459309 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-slash\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459352 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-slash" (OuterVolumeSpecName: "host-slash") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459429 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-kubelet\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459549 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459629 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-node-log\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459653 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-node-log" (OuterVolumeSpecName: "node-log") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459812 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459812 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459893 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459927 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459934 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-systemd-units\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459954 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.459992 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-ovn\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460051 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovn-node-metrics-cert\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460069 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-bin\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460096 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460102 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhn54\" (UniqueName: \"kubernetes.io/projected/28d6c67f-c4cd-4692-b490-b2b884c72db6-kube-api-access-bhn54\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460148 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-config\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460165 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460175 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-script-lib\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460210 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-netns\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460225 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-log-socket\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460250 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-netd\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460265 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-log-socket" (OuterVolumeSpecName: "log-socket") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460261 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460292 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-var-lib-openvswitch\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460306 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460318 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-env-overrides\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460327 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460357 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-etc-openvswitch\") pod \"28d6c67f-c4cd-4692-b490-b2b884c72db6\" (UID: \"28d6c67f-c4cd-4692-b490-b2b884c72db6\") " Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460585 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-node-log\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460618 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-env-overrides\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460678 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460700 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460733 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460739 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-slash\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460781 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460871 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460891 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.460956 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-ovn\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461056 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-systemd\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461110 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovnkube-config\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461143 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovnkube-script-lib\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461197 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461218 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srsk2\" (UniqueName: \"kubernetes.io/projected/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-kube-api-access-srsk2\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461233 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-kubelet\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461268 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-etc-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461321 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-cni-netd\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461417 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-run-netns\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461467 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-log-socket\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461547 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovn-node-metrics-cert\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461599 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-cni-bin\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461625 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-var-lib-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461650 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-systemd-units\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461745 4492 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461762 4492 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-node-log\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461774 4492 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461785 4492 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461793 4492 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461802 4492 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461810 4492 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461819 4492 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461830 4492 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-log-socket\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461839 4492 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461848 4492 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461858 4492 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461866 4492 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/28d6c67f-c4cd-4692-b490-b2b884c72db6-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461877 4492 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461885 4492 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461893 4492 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.461901 4492 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-host-slash\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.465371 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.465595 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28d6c67f-c4cd-4692-b490-b2b884c72db6-kube-api-access-bhn54" (OuterVolumeSpecName: "kube-api-access-bhn54") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "kube-api-access-bhn54". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.471726 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "28d6c67f-c4cd-4692-b490-b2b884c72db6" (UID: "28d6c67f-c4cd-4692-b490-b2b884c72db6"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562682 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovnkube-script-lib\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562735 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562764 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srsk2\" (UniqueName: \"kubernetes.io/projected/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-kube-api-access-srsk2\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562795 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-kubelet\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562815 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-etc-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562848 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-cni-netd\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562865 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562867 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-run-netns\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562906 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-kubelet\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.562923 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-run-netns\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563009 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-log-socket\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563037 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-cni-netd\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563076 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-etc-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563093 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovn-node-metrics-cert\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563147 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-log-socket\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563286 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-cni-bin\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563329 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-var-lib-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563365 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-systemd-units\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563398 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-cni-bin\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563418 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-var-lib-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563493 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-systemd-units\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563504 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-node-log\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563543 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-env-overrides\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563616 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-node-log\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563673 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563723 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563850 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-slash\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563874 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-ovn\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.563980 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-systemd\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564012 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovnkube-config\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564080 4492 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/28d6c67f-c4cd-4692-b490-b2b884c72db6-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564096 4492 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/28d6c67f-c4cd-4692-b490-b2b884c72db6-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564113 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhn54\" (UniqueName: \"kubernetes.io/projected/28d6c67f-c4cd-4692-b490-b2b884c72db6-kube-api-access-bhn54\") on node \"crc\" DevicePath \"\"" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564303 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-env-overrides\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564357 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-openvswitch\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564717 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovnkube-config\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.564760 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-run-ovn-kubernetes\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.565146 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-ovn\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.565151 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-host-slash\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.565221 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-run-systemd\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.565794 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovnkube-script-lib\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.568204 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-ovn-node-metrics-cert\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.582192 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srsk2\" (UniqueName: \"kubernetes.io/projected/7bbdff80-7c41-4118-8ae5-4106fa6aa1bd-kube-api-access-srsk2\") pod \"ovnkube-node-9pbsv\" (UID: \"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd\") " pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: I0220 06:49:25.698232 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:25 crc kubenswrapper[4492]: W0220 06:49:25.713851 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bbdff80_7c41_4118_8ae5_4106fa6aa1bd.slice/crio-d5c81db6be91c328dfd1dc14565066a19f04fada3ce8cda99de0085356acf7e6 WatchSource:0}: Error finding container d5c81db6be91c328dfd1dc14565066a19f04fada3ce8cda99de0085356acf7e6: Status 404 returned error can't find the container with id d5c81db6be91c328dfd1dc14565066a19f04fada3ce8cda99de0085356acf7e6 Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.290381 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/2.log" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.299096 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovn-acl-logging/0.log" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.299627 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f68mj_28d6c67f-c4cd-4692-b490-b2b884c72db6/ovn-controller/0.log" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.300837 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" event={"ID":"28d6c67f-c4cd-4692-b490-b2b884c72db6","Type":"ContainerDied","Data":"1b30d5ac2f119de9934e653a72f38b71d982caa643ae28dd8acede7d1b09843b"} Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.300897 4492 scope.go:117] "RemoveContainer" containerID="74a35c1868a572cbfcd7507555baf61f36d97d28af5ac21d481547bd4c44f779" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.301040 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f68mj" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.305276 4492 generic.go:334] "Generic (PLEG): container finished" podID="7bbdff80-7c41-4118-8ae5-4106fa6aa1bd" containerID="67ff38aea9011cff1946b2739c7d717ad6627b8702603912efadb39d377674d0" exitCode=0 Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.305314 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerDied","Data":"67ff38aea9011cff1946b2739c7d717ad6627b8702603912efadb39d377674d0"} Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.305341 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"d5c81db6be91c328dfd1dc14565066a19f04fada3ce8cda99de0085356acf7e6"} Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.338464 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f68mj"] Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.340190 4492 scope.go:117] "RemoveContainer" containerID="5ed893ab556c4a559961f8da46040674aed349e5973565edcb79f1e1e7b097b0" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.341881 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f68mj"] Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.359881 4492 scope.go:117] "RemoveContainer" containerID="3f7677b355b87bb14c95b7249de8425078f87106ccc8b86c3f95fe99004df1c3" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.381974 4492 scope.go:117] "RemoveContainer" containerID="8f3271988eb781c58ae282a05f7ed45586b4f6189a779bcb72baf76655e1610a" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.399433 4492 scope.go:117] "RemoveContainer" containerID="d4176f1b5c0d57bee185abeb855eb42cb62affe40b78849cbe44b3cecb6aac7c" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.410516 4492 scope.go:117] "RemoveContainer" containerID="5d2bbbe09e89466044fca186189f9d85b1245146b24347e961db7f61d2eee622" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.423387 4492 scope.go:117] "RemoveContainer" containerID="9a8689cc8d7cb80ae4a30d5852a39f53a90d13348a4bb1305db9acae17099c25" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.439055 4492 scope.go:117] "RemoveContainer" containerID="c683bd9c75a6b2e8ea32896f574025f4f49b07f031339394193272ef4952511a" Feb 20 06:49:26 crc kubenswrapper[4492]: I0220 06:49:26.458611 4492 scope.go:117] "RemoveContainer" containerID="478e9b26abd796027a31c93705109465c039f8158be6b9b25631493606bb7a87" Feb 20 06:49:27 crc kubenswrapper[4492]: I0220 06:49:27.316145 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"665ef7af314c36232bfef4c02cf13d1cbb02090e38f3c7e411541d3143c038d3"} Feb 20 06:49:27 crc kubenswrapper[4492]: I0220 06:49:27.316574 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"47e9b71cb04990aa8a9654f51b5cf0b6d915ff96903fb2d567f465d1ff22603b"} Feb 20 06:49:27 crc kubenswrapper[4492]: I0220 06:49:27.316594 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"bd4f1a53b325fc21019d10e6283d5cc71155fba3ca9c9a7688999721cd194e70"} Feb 20 06:49:27 crc kubenswrapper[4492]: I0220 06:49:27.316603 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"cd42be7f7cf35adbfc674475e57ee7527bbbe96850a2068a3121833610265557"} Feb 20 06:49:27 crc kubenswrapper[4492]: I0220 06:49:27.316611 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"3279fbbcdf63e27759ddec07517e0b84a472073e7244d84a5c2b101c3d746b1b"} Feb 20 06:49:27 crc kubenswrapper[4492]: I0220 06:49:27.316621 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"92d999abc1cbfa4e45ab138130cad1298866e3232af0a5328264178388db5886"} Feb 20 06:49:27 crc kubenswrapper[4492]: I0220 06:49:27.564867 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28d6c67f-c4cd-4692-b490-b2b884c72db6" path="/var/lib/kubelet/pods/28d6c67f-c4cd-4692-b490-b2b884c72db6/volumes" Feb 20 06:49:29 crc kubenswrapper[4492]: I0220 06:49:29.335364 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"dd1047fc7d4293de5577021751ea3a971ca34e04925bf9a206a624bca7cd0ac0"} Feb 20 06:49:31 crc kubenswrapper[4492]: I0220 06:49:31.356288 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" event={"ID":"7bbdff80-7c41-4118-8ae5-4106fa6aa1bd","Type":"ContainerStarted","Data":"e597c5868a7fdaf78a838e36f5be32deb8262db019c599a6b259998bb0e792af"} Feb 20 06:49:31 crc kubenswrapper[4492]: I0220 06:49:31.356866 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:31 crc kubenswrapper[4492]: I0220 06:49:31.356886 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:31 crc kubenswrapper[4492]: I0220 06:49:31.356899 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:31 crc kubenswrapper[4492]: I0220 06:49:31.388300 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" podStartSLOduration=6.388282373 podStartE2EDuration="6.388282373s" podCreationTimestamp="2026-02-20 06:49:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:49:31.387255748 +0000 UTC m=+528.158544726" watchObservedRunningTime="2026-02-20 06:49:31.388282373 +0000 UTC m=+528.159571351" Feb 20 06:49:31 crc kubenswrapper[4492]: I0220 06:49:31.391921 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:31 crc kubenswrapper[4492]: I0220 06:49:31.392419 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:35 crc kubenswrapper[4492]: I0220 06:49:35.556871 4492 scope.go:117] "RemoveContainer" containerID="e63ed096fe97531f09118ad4101afeac3694e124f9060047cedcde53421e659a" Feb 20 06:49:35 crc kubenswrapper[4492]: E0220 06:49:35.557433 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2rr9j_openshift-multus(dcc11973-022a-47f9-b1e2-23e945352e7f)\"" pod="openshift-multus/multus-2rr9j" podUID="dcc11973-022a-47f9-b1e2-23e945352e7f" Feb 20 06:49:39 crc kubenswrapper[4492]: I0220 06:49:39.311523 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:49:39 crc kubenswrapper[4492]: I0220 06:49:39.311933 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:49:46 crc kubenswrapper[4492]: I0220 06:49:46.557575 4492 scope.go:117] "RemoveContainer" containerID="e63ed096fe97531f09118ad4101afeac3694e124f9060047cedcde53421e659a" Feb 20 06:49:47 crc kubenswrapper[4492]: I0220 06:49:47.460196 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2rr9j_dcc11973-022a-47f9-b1e2-23e945352e7f/kube-multus/2.log" Feb 20 06:49:47 crc kubenswrapper[4492]: I0220 06:49:47.460721 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2rr9j" event={"ID":"dcc11973-022a-47f9-b1e2-23e945352e7f","Type":"ContainerStarted","Data":"8eb9cfe0d81707087ee3f94156e9194a78fb445f146e27573282ebe2b5a04ab6"} Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.408136 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg"] Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.409566 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.411631 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.419358 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg"] Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.535178 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.535232 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.535271 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dfrx\" (UniqueName: \"kubernetes.io/projected/634446d6-13a4-4a8d-9e28-b2abebca10a3-kube-api-access-2dfrx\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.636374 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dfrx\" (UniqueName: \"kubernetes.io/projected/634446d6-13a4-4a8d-9e28-b2abebca10a3-kube-api-access-2dfrx\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.636569 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.636603 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.637001 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-util\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.637104 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-bundle\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.654202 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dfrx\" (UniqueName: \"kubernetes.io/projected/634446d6-13a4-4a8d-9e28-b2abebca10a3-kube-api-access-2dfrx\") pod \"f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.717251 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9pbsv" Feb 20 06:49:55 crc kubenswrapper[4492]: I0220 06:49:55.723986 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:49:56 crc kubenswrapper[4492]: I0220 06:49:56.227840 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg"] Feb 20 06:49:56 crc kubenswrapper[4492]: I0220 06:49:56.515054 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" event={"ID":"634446d6-13a4-4a8d-9e28-b2abebca10a3","Type":"ContainerStarted","Data":"b2718df4e3f0d8ba3029bc7575938f226aecbb6a2155d528bf967edb227be6c1"} Feb 20 06:49:56 crc kubenswrapper[4492]: I0220 06:49:56.515128 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" event={"ID":"634446d6-13a4-4a8d-9e28-b2abebca10a3","Type":"ContainerStarted","Data":"14a8d9e6b9cacd2b2ad9254fb2e8c7b4b0bdb13acb3c3316bfe30f9b22d25d16"} Feb 20 06:49:57 crc kubenswrapper[4492]: I0220 06:49:57.522324 4492 generic.go:334] "Generic (PLEG): container finished" podID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerID="b2718df4e3f0d8ba3029bc7575938f226aecbb6a2155d528bf967edb227be6c1" exitCode=0 Feb 20 06:49:57 crc kubenswrapper[4492]: I0220 06:49:57.522413 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" event={"ID":"634446d6-13a4-4a8d-9e28-b2abebca10a3","Type":"ContainerDied","Data":"b2718df4e3f0d8ba3029bc7575938f226aecbb6a2155d528bf967edb227be6c1"} Feb 20 06:49:59 crc kubenswrapper[4492]: I0220 06:49:59.537539 4492 generic.go:334] "Generic (PLEG): container finished" podID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerID="c0a1540e9a2625be97ddac1e4a8bd3824f673f8299ee442ff9f4321363e19ac8" exitCode=0 Feb 20 06:49:59 crc kubenswrapper[4492]: I0220 06:49:59.537653 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" event={"ID":"634446d6-13a4-4a8d-9e28-b2abebca10a3","Type":"ContainerDied","Data":"c0a1540e9a2625be97ddac1e4a8bd3824f673f8299ee442ff9f4321363e19ac8"} Feb 20 06:50:00 crc kubenswrapper[4492]: I0220 06:50:00.547043 4492 generic.go:334] "Generic (PLEG): container finished" podID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerID="7d3c25673e6208743d4558dd9bbb7caa3cfe9d90f9a93774e434a397c44d01f9" exitCode=0 Feb 20 06:50:00 crc kubenswrapper[4492]: I0220 06:50:00.547097 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" event={"ID":"634446d6-13a4-4a8d-9e28-b2abebca10a3","Type":"ContainerDied","Data":"7d3c25673e6208743d4558dd9bbb7caa3cfe9d90f9a93774e434a397c44d01f9"} Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.733241 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.817111 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dfrx\" (UniqueName: \"kubernetes.io/projected/634446d6-13a4-4a8d-9e28-b2abebca10a3-kube-api-access-2dfrx\") pod \"634446d6-13a4-4a8d-9e28-b2abebca10a3\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.823365 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/634446d6-13a4-4a8d-9e28-b2abebca10a3-kube-api-access-2dfrx" (OuterVolumeSpecName: "kube-api-access-2dfrx") pod "634446d6-13a4-4a8d-9e28-b2abebca10a3" (UID: "634446d6-13a4-4a8d-9e28-b2abebca10a3"). InnerVolumeSpecName "kube-api-access-2dfrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.918145 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-util\") pod \"634446d6-13a4-4a8d-9e28-b2abebca10a3\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.918193 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-bundle\") pod \"634446d6-13a4-4a8d-9e28-b2abebca10a3\" (UID: \"634446d6-13a4-4a8d-9e28-b2abebca10a3\") " Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.918429 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dfrx\" (UniqueName: \"kubernetes.io/projected/634446d6-13a4-4a8d-9e28-b2abebca10a3-kube-api-access-2dfrx\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.918828 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-bundle" (OuterVolumeSpecName: "bundle") pod "634446d6-13a4-4a8d-9e28-b2abebca10a3" (UID: "634446d6-13a4-4a8d-9e28-b2abebca10a3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:50:01 crc kubenswrapper[4492]: I0220 06:50:01.926694 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-util" (OuterVolumeSpecName: "util") pod "634446d6-13a4-4a8d-9e28-b2abebca10a3" (UID: "634446d6-13a4-4a8d-9e28-b2abebca10a3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:50:02 crc kubenswrapper[4492]: I0220 06:50:02.019945 4492 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-util\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:02 crc kubenswrapper[4492]: I0220 06:50:02.019970 4492 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/634446d6-13a4-4a8d-9e28-b2abebca10a3-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:02 crc kubenswrapper[4492]: I0220 06:50:02.561601 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" event={"ID":"634446d6-13a4-4a8d-9e28-b2abebca10a3","Type":"ContainerDied","Data":"14a8d9e6b9cacd2b2ad9254fb2e8c7b4b0bdb13acb3c3316bfe30f9b22d25d16"} Feb 20 06:50:02 crc kubenswrapper[4492]: I0220 06:50:02.561990 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14a8d9e6b9cacd2b2ad9254fb2e8c7b4b0bdb13acb3c3316bfe30f9b22d25d16" Feb 20 06:50:02 crc kubenswrapper[4492]: I0220 06:50:02.561848 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.465554 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-4btwm"] Feb 20 06:50:04 crc kubenswrapper[4492]: E0220 06:50:04.466540 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerName="util" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.466644 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerName="util" Feb 20 06:50:04 crc kubenswrapper[4492]: E0220 06:50:04.466706 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerName="extract" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.466749 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerName="extract" Feb 20 06:50:04 crc kubenswrapper[4492]: E0220 06:50:04.466801 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerName="pull" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.466842 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerName="pull" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.467024 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="634446d6-13a4-4a8d-9e28-b2abebca10a3" containerName="extract" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.467569 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.469861 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.470150 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-tnm4j" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.473891 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.485519 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-4btwm"] Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.546068 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk78n\" (UniqueName: \"kubernetes.io/projected/026317d3-c00d-420e-b6cc-f5beba8f1e7f-kube-api-access-fk78n\") pod \"nmstate-operator-694c9596b7-4btwm\" (UID: \"026317d3-c00d-420e-b6cc-f5beba8f1e7f\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.648042 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk78n\" (UniqueName: \"kubernetes.io/projected/026317d3-c00d-420e-b6cc-f5beba8f1e7f-kube-api-access-fk78n\") pod \"nmstate-operator-694c9596b7-4btwm\" (UID: \"026317d3-c00d-420e-b6cc-f5beba8f1e7f\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.665781 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk78n\" (UniqueName: \"kubernetes.io/projected/026317d3-c00d-420e-b6cc-f5beba8f1e7f-kube-api-access-fk78n\") pod \"nmstate-operator-694c9596b7-4btwm\" (UID: \"026317d3-c00d-420e-b6cc-f5beba8f1e7f\") " pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.781744 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" Feb 20 06:50:04 crc kubenswrapper[4492]: I0220 06:50:04.962224 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-694c9596b7-4btwm"] Feb 20 06:50:05 crc kubenswrapper[4492]: I0220 06:50:05.577392 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" event={"ID":"026317d3-c00d-420e-b6cc-f5beba8f1e7f","Type":"ContainerStarted","Data":"9018d8991855cc71028ecd3dd7718ec6a9c59a272ebb33c74500cb6611bf616b"} Feb 20 06:50:07 crc kubenswrapper[4492]: I0220 06:50:07.598136 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" event={"ID":"026317d3-c00d-420e-b6cc-f5beba8f1e7f","Type":"ContainerStarted","Data":"3f164bb280aec9214954f5752615731ee2d4abf4bceea8f70b187743c2a246c5"} Feb 20 06:50:07 crc kubenswrapper[4492]: I0220 06:50:07.620877 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-694c9596b7-4btwm" podStartSLOduration=1.3248309950000001 podStartE2EDuration="3.620846209s" podCreationTimestamp="2026-02-20 06:50:04 +0000 UTC" firstStartedPulling="2026-02-20 06:50:04.976440404 +0000 UTC m=+561.747729382" lastFinishedPulling="2026-02-20 06:50:07.272455618 +0000 UTC m=+564.043744596" observedRunningTime="2026-02-20 06:50:07.61365943 +0000 UTC m=+564.384948408" watchObservedRunningTime="2026-02-20 06:50:07.620846209 +0000 UTC m=+564.392135188" Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.311008 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.311094 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.311152 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.311799 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"855da3473a6d9146a28853331dfc686fe2c8735ee6afb03897c4083de78fb717"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.311857 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://855da3473a6d9146a28853331dfc686fe2c8735ee6afb03897c4083de78fb717" gracePeriod=600 Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.613491 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="855da3473a6d9146a28853331dfc686fe2c8735ee6afb03897c4083de78fb717" exitCode=0 Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.613991 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"855da3473a6d9146a28853331dfc686fe2c8735ee6afb03897c4083de78fb717"} Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.614203 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"5090e1416f91468b7a37c5f3132c069b8613e7ce0e4c8c57bc77b7ef73cc9bce"} Feb 20 06:50:09 crc kubenswrapper[4492]: I0220 06:50:09.615649 4492 scope.go:117] "RemoveContainer" containerID="8f0bbbbef7d028667e57fa52189868db9e398ca3ba68113428d3adebbf972641" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.188408 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.189694 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.192216 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-kf2mm" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.212627 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.213376 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.216772 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsvvx\" (UniqueName: \"kubernetes.io/projected/0378fe77-c8df-4601-88af-0407d45e8d13-kube-api-access-fsvvx\") pod \"nmstate-metrics-58c85c668d-j8ddx\" (UID: \"0378fe77-c8df-4601-88af-0407d45e8d13\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.218945 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.229641 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.238030 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.269232 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-948rj"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.270075 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.317848 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsvvx\" (UniqueName: \"kubernetes.io/projected/0378fe77-c8df-4601-88af-0407d45e8d13-kube-api-access-fsvvx\") pod \"nmstate-metrics-58c85c668d-j8ddx\" (UID: \"0378fe77-c8df-4601-88af-0407d45e8d13\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.317897 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-nmstate-lock\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.317922 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-ovs-socket\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.318135 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/07c86583-b824-4484-9ddf-d4614fa1a99c-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tp9zd\" (UID: \"07c86583-b824-4484-9ddf-d4614fa1a99c\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.318201 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9xt5\" (UniqueName: \"kubernetes.io/projected/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-kube-api-access-q9xt5\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.318263 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-dbus-socket\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.318330 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdt8c\" (UniqueName: \"kubernetes.io/projected/07c86583-b824-4484-9ddf-d4614fa1a99c-kube-api-access-fdt8c\") pod \"nmstate-webhook-866bcb46dc-tp9zd\" (UID: \"07c86583-b824-4484-9ddf-d4614fa1a99c\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.350006 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsvvx\" (UniqueName: \"kubernetes.io/projected/0378fe77-c8df-4601-88af-0407d45e8d13-kube-api-access-fsvvx\") pod \"nmstate-metrics-58c85c668d-j8ddx\" (UID: \"0378fe77-c8df-4601-88af-0407d45e8d13\") " pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.362309 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.363599 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.367435 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.367839 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9dmx7" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.370339 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.371561 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.419781 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4rnl\" (UniqueName: \"kubernetes.io/projected/69627a87-2af9-40ee-860b-52cc31c8591d-kube-api-access-w4rnl\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.419846 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-nmstate-lock\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.419875 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-ovs-socket\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.419941 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/69627a87-2af9-40ee-860b-52cc31c8591d-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.419954 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-nmstate-lock\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.419990 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/69627a87-2af9-40ee-860b-52cc31c8591d-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.420041 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/07c86583-b824-4484-9ddf-d4614fa1a99c-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tp9zd\" (UID: \"07c86583-b824-4484-9ddf-d4614fa1a99c\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.420085 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9xt5\" (UniqueName: \"kubernetes.io/projected/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-kube-api-access-q9xt5\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.420093 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-ovs-socket\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.420134 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-dbus-socket\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: E0220 06:50:15.420175 4492 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Feb 20 06:50:15 crc kubenswrapper[4492]: E0220 06:50:15.420239 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07c86583-b824-4484-9ddf-d4614fa1a99c-tls-key-pair podName:07c86583-b824-4484-9ddf-d4614fa1a99c nodeName:}" failed. No retries permitted until 2026-02-20 06:50:15.920215069 +0000 UTC m=+572.691504048 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/07c86583-b824-4484-9ddf-d4614fa1a99c-tls-key-pair") pod "nmstate-webhook-866bcb46dc-tp9zd" (UID: "07c86583-b824-4484-9ddf-d4614fa1a99c") : secret "openshift-nmstate-webhook" not found Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.420181 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdt8c\" (UniqueName: \"kubernetes.io/projected/07c86583-b824-4484-9ddf-d4614fa1a99c-kube-api-access-fdt8c\") pod \"nmstate-webhook-866bcb46dc-tp9zd\" (UID: \"07c86583-b824-4484-9ddf-d4614fa1a99c\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.420802 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-dbus-socket\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.434864 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9xt5\" (UniqueName: \"kubernetes.io/projected/cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914-kube-api-access-q9xt5\") pod \"nmstate-handler-948rj\" (UID: \"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914\") " pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.435666 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdt8c\" (UniqueName: \"kubernetes.io/projected/07c86583-b824-4484-9ddf-d4614fa1a99c-kube-api-access-fdt8c\") pod \"nmstate-webhook-866bcb46dc-tp9zd\" (UID: \"07c86583-b824-4484-9ddf-d4614fa1a99c\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.508040 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.522297 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4rnl\" (UniqueName: \"kubernetes.io/projected/69627a87-2af9-40ee-860b-52cc31c8591d-kube-api-access-w4rnl\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.522549 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/69627a87-2af9-40ee-860b-52cc31c8591d-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.522661 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/69627a87-2af9-40ee-860b-52cc31c8591d-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.523647 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/69627a87-2af9-40ee-860b-52cc31c8591d-nginx-conf\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.528387 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/69627a87-2af9-40ee-860b-52cc31c8591d-plugin-serving-cert\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.548076 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4rnl\" (UniqueName: \"kubernetes.io/projected/69627a87-2af9-40ee-860b-52cc31c8591d-kube-api-access-w4rnl\") pod \"nmstate-console-plugin-5c78fc5d65-2vhlz\" (UID: \"69627a87-2af9-40ee-860b-52cc31c8591d\") " pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.582030 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.598271 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-54b4fd76b5-ftpjc"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.599156 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.619265 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54b4fd76b5-ftpjc"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.625816 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-serving-cert\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.625876 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-oauth-config\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.625912 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-trusted-ca-bundle\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.625936 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-service-ca\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.625989 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-oauth-serving-cert\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.626021 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-config\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.626047 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hztvw\" (UniqueName: \"kubernetes.io/projected/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-kube-api-access-hztvw\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.658799 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-948rj" event={"ID":"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914","Type":"ContainerStarted","Data":"01d1784cd9ae9748ba9f42d0baa7175bb57afd553308a5611339981c0aeb4ca9"} Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.679811 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.726882 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-service-ca\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.726945 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-oauth-serving-cert\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.726977 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-config\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.727007 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hztvw\" (UniqueName: \"kubernetes.io/projected/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-kube-api-access-hztvw\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.727031 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-serving-cert\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.727061 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-oauth-config\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.727093 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-trusted-ca-bundle\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.727965 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-oauth-serving-cert\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.728128 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-trusted-ca-bundle\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.728782 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-service-ca\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.732284 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-serving-cert\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.736025 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-config\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.739228 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-console-oauth-config\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.744564 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx"] Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.746867 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hztvw\" (UniqueName: \"kubernetes.io/projected/cb9bc1ac-50b8-4b0e-b063-cd10d29af261-kube-api-access-hztvw\") pod \"console-54b4fd76b5-ftpjc\" (UID: \"cb9bc1ac-50b8-4b0e-b063-cd10d29af261\") " pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: W0220 06:50:15.749533 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0378fe77_c8df_4601_88af_0407d45e8d13.slice/crio-35687717956d4e21754a28353fc6efd05222d55ee969733b820a601053e6ef77 WatchSource:0}: Error finding container 35687717956d4e21754a28353fc6efd05222d55ee969733b820a601053e6ef77: Status 404 returned error can't find the container with id 35687717956d4e21754a28353fc6efd05222d55ee969733b820a601053e6ef77 Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.863633 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz"] Feb 20 06:50:15 crc kubenswrapper[4492]: W0220 06:50:15.867274 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69627a87_2af9_40ee_860b_52cc31c8591d.slice/crio-928177da39a9edda7d6d12e874d75f8124ecc876ab18380d5dd857c11977aed9 WatchSource:0}: Error finding container 928177da39a9edda7d6d12e874d75f8124ecc876ab18380d5dd857c11977aed9: Status 404 returned error can't find the container with id 928177da39a9edda7d6d12e874d75f8124ecc876ab18380d5dd857c11977aed9 Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.919379 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.930341 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/07c86583-b824-4484-9ddf-d4614fa1a99c-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tp9zd\" (UID: \"07c86583-b824-4484-9ddf-d4614fa1a99c\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:15 crc kubenswrapper[4492]: I0220 06:50:15.934664 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/07c86583-b824-4484-9ddf-d4614fa1a99c-tls-key-pair\") pod \"nmstate-webhook-866bcb46dc-tp9zd\" (UID: \"07c86583-b824-4484-9ddf-d4614fa1a99c\") " pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.103428 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-54b4fd76b5-ftpjc"] Feb 20 06:50:16 crc kubenswrapper[4492]: W0220 06:50:16.111954 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb9bc1ac_50b8_4b0e_b063_cd10d29af261.slice/crio-88711e8ebbf9484aa660c3134154f45a965a38f25d74e75f62cb117a8f483174 WatchSource:0}: Error finding container 88711e8ebbf9484aa660c3134154f45a965a38f25d74e75f62cb117a8f483174: Status 404 returned error can't find the container with id 88711e8ebbf9484aa660c3134154f45a965a38f25d74e75f62cb117a8f483174 Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.126252 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.317917 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd"] Feb 20 06:50:16 crc kubenswrapper[4492]: W0220 06:50:16.320373 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c86583_b824_4484_9ddf_d4614fa1a99c.slice/crio-efc0f41bdee05475384222015a2b8f10d33cd99b5b0b61ae19f8fbe3473c1b2d WatchSource:0}: Error finding container efc0f41bdee05475384222015a2b8f10d33cd99b5b0b61ae19f8fbe3473c1b2d: Status 404 returned error can't find the container with id efc0f41bdee05475384222015a2b8f10d33cd99b5b0b61ae19f8fbe3473c1b2d Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.666808 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" event={"ID":"69627a87-2af9-40ee-860b-52cc31c8591d","Type":"ContainerStarted","Data":"928177da39a9edda7d6d12e874d75f8124ecc876ab18380d5dd857c11977aed9"} Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.668721 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54b4fd76b5-ftpjc" event={"ID":"cb9bc1ac-50b8-4b0e-b063-cd10d29af261","Type":"ContainerStarted","Data":"24f9261d58ef0f1b4184b3a909df3d7659017d905b4ecaf3bcbddda581df94c7"} Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.668766 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-54b4fd76b5-ftpjc" event={"ID":"cb9bc1ac-50b8-4b0e-b063-cd10d29af261","Type":"ContainerStarted","Data":"88711e8ebbf9484aa660c3134154f45a965a38f25d74e75f62cb117a8f483174"} Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.669785 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" event={"ID":"07c86583-b824-4484-9ddf-d4614fa1a99c","Type":"ContainerStarted","Data":"efc0f41bdee05475384222015a2b8f10d33cd99b5b0b61ae19f8fbe3473c1b2d"} Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.671109 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" event={"ID":"0378fe77-c8df-4601-88af-0407d45e8d13","Type":"ContainerStarted","Data":"35687717956d4e21754a28353fc6efd05222d55ee969733b820a601053e6ef77"} Feb 20 06:50:16 crc kubenswrapper[4492]: I0220 06:50:16.697750 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-54b4fd76b5-ftpjc" podStartSLOduration=1.6977244379999998 podStartE2EDuration="1.697724438s" podCreationTimestamp="2026-02-20 06:50:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:50:16.693807148 +0000 UTC m=+573.465096127" watchObservedRunningTime="2026-02-20 06:50:16.697724438 +0000 UTC m=+573.469013416" Feb 20 06:50:18 crc kubenswrapper[4492]: I0220 06:50:18.686090 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" event={"ID":"07c86583-b824-4484-9ddf-d4614fa1a99c","Type":"ContainerStarted","Data":"d422f8cffe953b85e4323ef6b3cd31835e547b11d28f4d567df33dd70376ff0c"} Feb 20 06:50:18 crc kubenswrapper[4492]: I0220 06:50:18.686541 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:18 crc kubenswrapper[4492]: I0220 06:50:18.688889 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" event={"ID":"69627a87-2af9-40ee-860b-52cc31c8591d","Type":"ContainerStarted","Data":"ac17662bb8e34852e68234f5f2216c6eabf52ac21f9365fb55f98a89b6c0ebaf"} Feb 20 06:50:18 crc kubenswrapper[4492]: I0220 06:50:18.700306 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" podStartSLOduration=1.540416203 podStartE2EDuration="3.700288448s" podCreationTimestamp="2026-02-20 06:50:15 +0000 UTC" firstStartedPulling="2026-02-20 06:50:16.322983502 +0000 UTC m=+573.094272480" lastFinishedPulling="2026-02-20 06:50:18.482855746 +0000 UTC m=+575.254144725" observedRunningTime="2026-02-20 06:50:18.698436448 +0000 UTC m=+575.469725427" watchObservedRunningTime="2026-02-20 06:50:18.700288448 +0000 UTC m=+575.471577426" Feb 20 06:50:18 crc kubenswrapper[4492]: I0220 06:50:18.711931 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5c78fc5d65-2vhlz" podStartSLOduration=1.111076189 podStartE2EDuration="3.711908511s" podCreationTimestamp="2026-02-20 06:50:15 +0000 UTC" firstStartedPulling="2026-02-20 06:50:15.871452659 +0000 UTC m=+572.642741637" lastFinishedPulling="2026-02-20 06:50:18.472284981 +0000 UTC m=+575.243573959" observedRunningTime="2026-02-20 06:50:18.711164018 +0000 UTC m=+575.482452996" watchObservedRunningTime="2026-02-20 06:50:18.711908511 +0000 UTC m=+575.483197489" Feb 20 06:50:19 crc kubenswrapper[4492]: I0220 06:50:19.696899 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-948rj" event={"ID":"cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914","Type":"ContainerStarted","Data":"bfae010a2cad7d1e53a6ea404dfc08e0e03c3d2ea16f1b27a54cb85133773650"} Feb 20 06:50:19 crc kubenswrapper[4492]: I0220 06:50:19.697901 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:19 crc kubenswrapper[4492]: I0220 06:50:19.720080 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-948rj" podStartSLOduration=1.86016336 podStartE2EDuration="4.720059768s" podCreationTimestamp="2026-02-20 06:50:15 +0000 UTC" firstStartedPulling="2026-02-20 06:50:15.614407126 +0000 UTC m=+572.385696104" lastFinishedPulling="2026-02-20 06:50:18.474303533 +0000 UTC m=+575.245592512" observedRunningTime="2026-02-20 06:50:19.718743678 +0000 UTC m=+576.490032656" watchObservedRunningTime="2026-02-20 06:50:19.720059768 +0000 UTC m=+576.491348747" Feb 20 06:50:20 crc kubenswrapper[4492]: I0220 06:50:20.704693 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" event={"ID":"0378fe77-c8df-4601-88af-0407d45e8d13","Type":"ContainerStarted","Data":"1554e18950bdfac6d4aa15e925c3884d09e2bfba43751eaaedfa441f7a7a5328"} Feb 20 06:50:22 crc kubenswrapper[4492]: I0220 06:50:22.724795 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" event={"ID":"0378fe77-c8df-4601-88af-0407d45e8d13","Type":"ContainerStarted","Data":"ad9cdd878fc165790e69b0310145046c7cb90e5654ad92fce4ba0c2b1143add8"} Feb 20 06:50:22 crc kubenswrapper[4492]: I0220 06:50:22.744810 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58c85c668d-j8ddx" podStartSLOduration=1.2883056960000001 podStartE2EDuration="7.744789619s" podCreationTimestamp="2026-02-20 06:50:15 +0000 UTC" firstStartedPulling="2026-02-20 06:50:15.752068345 +0000 UTC m=+572.523357322" lastFinishedPulling="2026-02-20 06:50:22.208552266 +0000 UTC m=+578.979841245" observedRunningTime="2026-02-20 06:50:22.741901829 +0000 UTC m=+579.513190806" watchObservedRunningTime="2026-02-20 06:50:22.744789619 +0000 UTC m=+579.516078587" Feb 20 06:50:25 crc kubenswrapper[4492]: I0220 06:50:25.607844 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-948rj" Feb 20 06:50:25 crc kubenswrapper[4492]: I0220 06:50:25.919746 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:25 crc kubenswrapper[4492]: I0220 06:50:25.919818 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:25 crc kubenswrapper[4492]: I0220 06:50:25.924341 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:26 crc kubenswrapper[4492]: I0220 06:50:26.755272 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-54b4fd76b5-ftpjc" Feb 20 06:50:26 crc kubenswrapper[4492]: I0220 06:50:26.814286 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wbkxj"] Feb 20 06:50:36 crc kubenswrapper[4492]: I0220 06:50:36.132152 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-866bcb46dc-tp9zd" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.227284 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q"] Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.229387 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.232419 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.242110 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q"] Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.253209 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.253279 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95ptj\" (UniqueName: \"kubernetes.io/projected/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-kube-api-access-95ptj\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.253404 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.355008 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.355076 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.355122 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95ptj\" (UniqueName: \"kubernetes.io/projected/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-kube-api-access-95ptj\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.355644 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.355718 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.374897 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95ptj\" (UniqueName: \"kubernetes.io/projected/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-kube-api-access-95ptj\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.555934 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:48 crc kubenswrapper[4492]: I0220 06:50:48.952905 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q"] Feb 20 06:50:49 crc kubenswrapper[4492]: I0220 06:50:49.915293 4492 generic.go:334] "Generic (PLEG): container finished" podID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerID="f3fec95680e7fa397102eb8336becffd233c449d89516d806fe416918fcd45e9" exitCode=0 Feb 20 06:50:49 crc kubenswrapper[4492]: I0220 06:50:49.915369 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" event={"ID":"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c","Type":"ContainerDied","Data":"f3fec95680e7fa397102eb8336becffd233c449d89516d806fe416918fcd45e9"} Feb 20 06:50:49 crc kubenswrapper[4492]: I0220 06:50:49.915423 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" event={"ID":"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c","Type":"ContainerStarted","Data":"fc60c2eb74f6119a8db34c9567684433b24c2e4c969a3ba7ed8f8f2eb4682ddb"} Feb 20 06:50:51 crc kubenswrapper[4492]: I0220 06:50:51.844289 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-wbkxj" podUID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" containerName="console" containerID="cri-o://914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b" gracePeriod=15 Feb 20 06:50:51 crc kubenswrapper[4492]: I0220 06:50:51.931372 4492 generic.go:334] "Generic (PLEG): container finished" podID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerID="3c94a7cbe8d0a1ac4322254c3b542c645304453232e4747b244b4baa630b2428" exitCode=0 Feb 20 06:50:51 crc kubenswrapper[4492]: I0220 06:50:51.931431 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" event={"ID":"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c","Type":"ContainerDied","Data":"3c94a7cbe8d0a1ac4322254c3b542c645304453232e4747b244b4baa630b2428"} Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.161454 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wbkxj_b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb/console/0.log" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.161839 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.201653 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-serving-cert\") pod \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.201760 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-trusted-ca-bundle\") pod \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.202575 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-service-ca" (OuterVolumeSpecName: "service-ca") pod "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" (UID: "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.202637 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" (UID: "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.201796 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-service-ca\") pod \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.202735 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-config\") pod \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.202804 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8rhf\" (UniqueName: \"kubernetes.io/projected/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-kube-api-access-b8rhf\") pod \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.202844 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-oauth-config\") pod \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.202876 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-oauth-serving-cert\") pod \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\" (UID: \"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb\") " Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.203192 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-config" (OuterVolumeSpecName: "console-config") pod "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" (UID: "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.203298 4492 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.203325 4492 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-service-ca\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.203337 4492 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.203526 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" (UID: "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.209295 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" (UID: "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.209346 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-kube-api-access-b8rhf" (OuterVolumeSpecName: "kube-api-access-b8rhf") pod "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" (UID: "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb"). InnerVolumeSpecName "kube-api-access-b8rhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.209545 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" (UID: "b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.305068 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8rhf\" (UniqueName: \"kubernetes.io/projected/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-kube-api-access-b8rhf\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.305106 4492 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.305118 4492 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.305128 4492 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.938222 4492 generic.go:334] "Generic (PLEG): container finished" podID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerID="3e560ea3c8a7e7b2f7de1ad04ed6f4046774c399487e864a01f802f3544ed977" exitCode=0 Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.938282 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" event={"ID":"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c","Type":"ContainerDied","Data":"3e560ea3c8a7e7b2f7de1ad04ed6f4046774c399487e864a01f802f3544ed977"} Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.940335 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wbkxj_b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb/console/0.log" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.940386 4492 generic.go:334] "Generic (PLEG): container finished" podID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" containerID="914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b" exitCode=2 Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.940422 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wbkxj" event={"ID":"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb","Type":"ContainerDied","Data":"914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b"} Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.940453 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wbkxj" event={"ID":"b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb","Type":"ContainerDied","Data":"01b2c9e52346030596d67613c35a98b9e1f9b177acf86f7ff60dc52d66d18135"} Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.940495 4492 scope.go:117] "RemoveContainer" containerID="914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.940607 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wbkxj" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.956006 4492 scope.go:117] "RemoveContainer" containerID="914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b" Feb 20 06:50:52 crc kubenswrapper[4492]: E0220 06:50:52.956318 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b\": container with ID starting with 914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b not found: ID does not exist" containerID="914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.956407 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b"} err="failed to get container status \"914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b\": rpc error: code = NotFound desc = could not find container \"914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b\": container with ID starting with 914108f6585f72e416a58a7ceba269f89c36179ee619756de2143ba30b431a7b not found: ID does not exist" Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.964963 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wbkxj"] Feb 20 06:50:52 crc kubenswrapper[4492]: I0220 06:50:52.971155 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-wbkxj"] Feb 20 06:50:53 crc kubenswrapper[4492]: I0220 06:50:53.564033 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" path="/var/lib/kubelet/pods/b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb/volumes" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.149346 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.330411 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95ptj\" (UniqueName: \"kubernetes.io/projected/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-kube-api-access-95ptj\") pod \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.330526 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-bundle\") pod \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.330603 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-util\") pod \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\" (UID: \"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c\") " Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.331590 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-bundle" (OuterVolumeSpecName: "bundle") pod "d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" (UID: "d4e34ff6-7469-44df-8ab0-fdb3f6d0535c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.337212 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-kube-api-access-95ptj" (OuterVolumeSpecName: "kube-api-access-95ptj") pod "d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" (UID: "d4e34ff6-7469-44df-8ab0-fdb3f6d0535c"). InnerVolumeSpecName "kube-api-access-95ptj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.341811 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-util" (OuterVolumeSpecName: "util") pod "d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" (UID: "d4e34ff6-7469-44df-8ab0-fdb3f6d0535c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.432441 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95ptj\" (UniqueName: \"kubernetes.io/projected/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-kube-api-access-95ptj\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.432505 4492 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.432518 4492 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4e34ff6-7469-44df-8ab0-fdb3f6d0535c-util\") on node \"crc\" DevicePath \"\"" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.956460 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" event={"ID":"d4e34ff6-7469-44df-8ab0-fdb3f6d0535c","Type":"ContainerDied","Data":"fc60c2eb74f6119a8db34c9567684433b24c2e4c969a3ba7ed8f8f2eb4682ddb"} Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.956548 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc60c2eb74f6119a8db34c9567684433b24c2e4c969a3ba7ed8f8f2eb4682ddb" Feb 20 06:50:54 crc kubenswrapper[4492]: I0220 06:50:54.956568 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.336822 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b"] Feb 20 06:51:03 crc kubenswrapper[4492]: E0220 06:51:03.337241 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerName="util" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.337255 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerName="util" Feb 20 06:51:03 crc kubenswrapper[4492]: E0220 06:51:03.337265 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" containerName="console" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.337271 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" containerName="console" Feb 20 06:51:03 crc kubenswrapper[4492]: E0220 06:51:03.337281 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerName="extract" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.337287 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerName="extract" Feb 20 06:51:03 crc kubenswrapper[4492]: E0220 06:51:03.337293 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerName="pull" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.337299 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerName="pull" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.337389 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4e34ff6-7469-44df-8ab0-fdb3f6d0535c" containerName="extract" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.337404 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3f90beb-ffe5-460a-85ce-f3e78fc6f9bb" containerName="console" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.337784 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.339366 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-97tjl" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.339391 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.339700 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.340174 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.340222 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.350408 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fe223a3-3ace-4d01-8379-d74032c024a0-webhook-cert\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.350516 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fe223a3-3ace-4d01-8379-d74032c024a0-apiservice-cert\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.350547 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsp4s\" (UniqueName: \"kubernetes.io/projected/2fe223a3-3ace-4d01-8379-d74032c024a0-kube-api-access-jsp4s\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.356379 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b"] Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.451852 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fe223a3-3ace-4d01-8379-d74032c024a0-apiservice-cert\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.451894 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsp4s\" (UniqueName: \"kubernetes.io/projected/2fe223a3-3ace-4d01-8379-d74032c024a0-kube-api-access-jsp4s\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.451932 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fe223a3-3ace-4d01-8379-d74032c024a0-webhook-cert\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.458385 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2fe223a3-3ace-4d01-8379-d74032c024a0-apiservice-cert\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.470108 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2fe223a3-3ace-4d01-8379-d74032c024a0-webhook-cert\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.490341 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsp4s\" (UniqueName: \"kubernetes.io/projected/2fe223a3-3ace-4d01-8379-d74032c024a0-kube-api-access-jsp4s\") pod \"metallb-operator-controller-manager-76ffddff97-9dv5b\" (UID: \"2fe223a3-3ace-4d01-8379-d74032c024a0\") " pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.653442 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.752981 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v"] Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.753785 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.758629 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vczkd" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.764829 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.764922 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.785817 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v"] Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.856549 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5f5t\" (UniqueName: \"kubernetes.io/projected/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-kube-api-access-s5f5t\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.856804 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-webhook-cert\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.856951 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-apiservice-cert\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.960284 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-webhook-cert\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.960562 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-apiservice-cert\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.960673 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5f5t\" (UniqueName: \"kubernetes.io/projected/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-kube-api-access-s5f5t\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.966413 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-apiservice-cert\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.967122 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-webhook-cert\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:03 crc kubenswrapper[4492]: I0220 06:51:03.977964 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5f5t\" (UniqueName: \"kubernetes.io/projected/9ef7522d-21b4-4acf-9cf3-c0217fcc173b-kube-api-access-s5f5t\") pod \"metallb-operator-webhook-server-7d989b6fcf-r2r6v\" (UID: \"9ef7522d-21b4-4acf-9cf3-c0217fcc173b\") " pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:04 crc kubenswrapper[4492]: I0220 06:51:04.067068 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:04 crc kubenswrapper[4492]: I0220 06:51:04.256365 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b"] Feb 20 06:51:04 crc kubenswrapper[4492]: I0220 06:51:04.275287 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v"] Feb 20 06:51:04 crc kubenswrapper[4492]: W0220 06:51:04.280880 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ef7522d_21b4_4acf_9cf3_c0217fcc173b.slice/crio-b54058a6681db60ef09c7b73d1eafeb126aa548b359c88dfd6ead14b50dcb996 WatchSource:0}: Error finding container b54058a6681db60ef09c7b73d1eafeb126aa548b359c88dfd6ead14b50dcb996: Status 404 returned error can't find the container with id b54058a6681db60ef09c7b73d1eafeb126aa548b359c88dfd6ead14b50dcb996 Feb 20 06:51:05 crc kubenswrapper[4492]: I0220 06:51:05.015379 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" event={"ID":"2fe223a3-3ace-4d01-8379-d74032c024a0","Type":"ContainerStarted","Data":"31a6bbe979f662be5ceb84ad4cef886c186a21ae170bd161ea743ade5e7c57e2"} Feb 20 06:51:05 crc kubenswrapper[4492]: I0220 06:51:05.017281 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" event={"ID":"9ef7522d-21b4-4acf-9cf3-c0217fcc173b","Type":"ContainerStarted","Data":"b54058a6681db60ef09c7b73d1eafeb126aa548b359c88dfd6ead14b50dcb996"} Feb 20 06:51:10 crc kubenswrapper[4492]: I0220 06:51:10.094261 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" event={"ID":"2fe223a3-3ace-4d01-8379-d74032c024a0","Type":"ContainerStarted","Data":"55058b8feb49f621bec3c874bfaf2c073ed0939e644dce0bc8260223a19c6e00"} Feb 20 06:51:10 crc kubenswrapper[4492]: I0220 06:51:10.095127 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:10 crc kubenswrapper[4492]: I0220 06:51:10.097567 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" event={"ID":"9ef7522d-21b4-4acf-9cf3-c0217fcc173b","Type":"ContainerStarted","Data":"cbc834eb8ad60fbaa3ee979f0f825ad04d25e6db21b4268d55453e139d4b7b85"} Feb 20 06:51:10 crc kubenswrapper[4492]: I0220 06:51:10.098400 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:10 crc kubenswrapper[4492]: I0220 06:51:10.120228 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" podStartSLOduration=3.670453927 podStartE2EDuration="7.120204509s" podCreationTimestamp="2026-02-20 06:51:03 +0000 UTC" firstStartedPulling="2026-02-20 06:51:04.271839929 +0000 UTC m=+621.043128907" lastFinishedPulling="2026-02-20 06:51:07.721590511 +0000 UTC m=+624.492879489" observedRunningTime="2026-02-20 06:51:10.117324128 +0000 UTC m=+626.888613106" watchObservedRunningTime="2026-02-20 06:51:10.120204509 +0000 UTC m=+626.891493477" Feb 20 06:51:10 crc kubenswrapper[4492]: I0220 06:51:10.141393 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" podStartSLOduration=2.280284932 podStartE2EDuration="7.141375144s" podCreationTimestamp="2026-02-20 06:51:03 +0000 UTC" firstStartedPulling="2026-02-20 06:51:04.284769626 +0000 UTC m=+621.056058605" lastFinishedPulling="2026-02-20 06:51:09.145859839 +0000 UTC m=+625.917148817" observedRunningTime="2026-02-20 06:51:10.135784012 +0000 UTC m=+626.907072991" watchObservedRunningTime="2026-02-20 06:51:10.141375144 +0000 UTC m=+626.912664122" Feb 20 06:51:24 crc kubenswrapper[4492]: I0220 06:51:24.107136 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7d989b6fcf-r2r6v" Feb 20 06:51:43 crc kubenswrapper[4492]: I0220 06:51:43.655966 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-76ffddff97-9dv5b" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.199464 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-sps8s"] Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.201547 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.205024 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.205030 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.206791 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6t8rw" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.212571 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr"] Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.213493 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.217509 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.226601 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr"] Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227580 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-frr-sockets\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227615 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n944\" (UniqueName: \"kubernetes.io/projected/1224d7af-027f-4a22-b675-606a9650f716-kube-api-access-9n944\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227641 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnsch\" (UniqueName: \"kubernetes.io/projected/c3b6f338-04d7-40c8-baa6-8f17a228a458-kube-api-access-pnsch\") pod \"frr-k8s-webhook-server-78b44bf5bb-lmpxr\" (UID: \"c3b6f338-04d7-40c8-baa6-8f17a228a458\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227667 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1224d7af-027f-4a22-b675-606a9650f716-metrics-certs\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227692 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-reloader\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227714 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-metrics\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227738 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1224d7af-027f-4a22-b675-606a9650f716-frr-startup\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227765 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-frr-conf\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.227780 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3b6f338-04d7-40c8-baa6-8f17a228a458-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-lmpxr\" (UID: \"c3b6f338-04d7-40c8-baa6-8f17a228a458\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.296066 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-rnqvv"] Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.297034 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.298955 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.298986 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.299176 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kgbc7" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.300067 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.308127 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-28vtj"] Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.309920 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.311163 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.323005 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-28vtj"] Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.329038 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1224d7af-027f-4a22-b675-606a9650f716-frr-startup\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.329181 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-frr-conf\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.329257 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3b6f338-04d7-40c8-baa6-8f17a228a458-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-lmpxr\" (UID: \"c3b6f338-04d7-40c8-baa6-8f17a228a458\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.329930 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-frr-sockets\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.330318 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n944\" (UniqueName: \"kubernetes.io/projected/1224d7af-027f-4a22-b675-606a9650f716-kube-api-access-9n944\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.330641 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnsch\" (UniqueName: \"kubernetes.io/projected/c3b6f338-04d7-40c8-baa6-8f17a228a458-kube-api-access-pnsch\") pod \"frr-k8s-webhook-server-78b44bf5bb-lmpxr\" (UID: \"c3b6f338-04d7-40c8-baa6-8f17a228a458\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.330753 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1224d7af-027f-4a22-b675-606a9650f716-metrics-certs\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.330846 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-reloader\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.330925 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-metrics\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.331181 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-metrics\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.330245 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1224d7af-027f-4a22-b675-606a9650f716-frr-startup\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.330274 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-frr-sockets\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.339197 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c3b6f338-04d7-40c8-baa6-8f17a228a458-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-lmpxr\" (UID: \"c3b6f338-04d7-40c8-baa6-8f17a228a458\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.329786 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-frr-conf\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.339505 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1224d7af-027f-4a22-b675-606a9650f716-reloader\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.345747 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1224d7af-027f-4a22-b675-606a9650f716-metrics-certs\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.371052 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnsch\" (UniqueName: \"kubernetes.io/projected/c3b6f338-04d7-40c8-baa6-8f17a228a458-kube-api-access-pnsch\") pod \"frr-k8s-webhook-server-78b44bf5bb-lmpxr\" (UID: \"c3b6f338-04d7-40c8-baa6-8f17a228a458\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.387210 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n944\" (UniqueName: \"kubernetes.io/projected/1224d7af-027f-4a22-b675-606a9650f716-kube-api-access-9n944\") pod \"frr-k8s-sps8s\" (UID: \"1224d7af-027f-4a22-b675-606a9650f716\") " pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.431717 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.431794 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-metrics-certs\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.431820 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-cert\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.431845 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7837488e-131f-4866-9417-5c9605902130-metallb-excludel2\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.431878 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxqng\" (UniqueName: \"kubernetes.io/projected/7837488e-131f-4866-9417-5c9605902130-kube-api-access-cxqng\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.431914 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjvsq\" (UniqueName: \"kubernetes.io/projected/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-kube-api-access-pjvsq\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.431934 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-metrics-certs\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.523593 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.535945 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.546720 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxqng\" (UniqueName: \"kubernetes.io/projected/7837488e-131f-4866-9417-5c9605902130-kube-api-access-cxqng\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.546795 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjvsq\" (UniqueName: \"kubernetes.io/projected/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-kube-api-access-pjvsq\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.546820 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-metrics-certs\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.546864 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.546927 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-metrics-certs\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.546946 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-cert\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.546971 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7837488e-131f-4866-9417-5c9605902130-metallb-excludel2\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: E0220 06:51:44.547155 4492 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 20 06:51:44 crc kubenswrapper[4492]: E0220 06:51:44.547266 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist podName:7837488e-131f-4866-9417-5c9605902130 nodeName:}" failed. No retries permitted until 2026-02-20 06:51:45.047248339 +0000 UTC m=+661.818537317 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist") pod "speaker-rnqvv" (UID: "7837488e-131f-4866-9417-5c9605902130") : secret "metallb-memberlist" not found Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.547593 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7837488e-131f-4866-9417-5c9605902130-metallb-excludel2\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: E0220 06:51:44.547729 4492 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Feb 20 06:51:44 crc kubenswrapper[4492]: E0220 06:51:44.547816 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-metrics-certs podName:6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3 nodeName:}" failed. No retries permitted until 2026-02-20 06:51:45.04780575 +0000 UTC m=+661.819094727 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-metrics-certs") pod "controller-69bbfbf88f-28vtj" (UID: "6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3") : secret "controller-certs-secret" not found Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.550433 4492 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.550535 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-metrics-certs\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.565100 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-cert\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.569572 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjvsq\" (UniqueName: \"kubernetes.io/projected/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-kube-api-access-pjvsq\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.577292 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxqng\" (UniqueName: \"kubernetes.io/projected/7837488e-131f-4866-9417-5c9605902130-kube-api-access-cxqng\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:44 crc kubenswrapper[4492]: I0220 06:51:44.745581 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr"] Feb 20 06:51:44 crc kubenswrapper[4492]: W0220 06:51:44.748714 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3b6f338_04d7_40c8_baa6_8f17a228a458.slice/crio-5537f7069b60a13c560d49d54b16ecf99cc41a1342f366e44c39e3093c92b627 WatchSource:0}: Error finding container 5537f7069b60a13c560d49d54b16ecf99cc41a1342f366e44c39e3093c92b627: Status 404 returned error can't find the container with id 5537f7069b60a13c560d49d54b16ecf99cc41a1342f366e44c39e3093c92b627 Feb 20 06:51:45 crc kubenswrapper[4492]: I0220 06:51:45.056405 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:45 crc kubenswrapper[4492]: I0220 06:51:45.056588 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-metrics-certs\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:45 crc kubenswrapper[4492]: E0220 06:51:45.056599 4492 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 20 06:51:45 crc kubenswrapper[4492]: E0220 06:51:45.056679 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist podName:7837488e-131f-4866-9417-5c9605902130 nodeName:}" failed. No retries permitted until 2026-02-20 06:51:46.056661446 +0000 UTC m=+662.827950424 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist") pod "speaker-rnqvv" (UID: "7837488e-131f-4866-9417-5c9605902130") : secret "metallb-memberlist" not found Feb 20 06:51:45 crc kubenswrapper[4492]: I0220 06:51:45.061005 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3-metrics-certs\") pod \"controller-69bbfbf88f-28vtj\" (UID: \"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3\") " pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:45 crc kubenswrapper[4492]: I0220 06:51:45.221596 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:45 crc kubenswrapper[4492]: I0220 06:51:45.317726 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerStarted","Data":"88e39c604db137c05b03ab641b4e5c5af17a7b0cb3732ef2e825c42799f7c8a1"} Feb 20 06:51:45 crc kubenswrapper[4492]: I0220 06:51:45.319367 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" event={"ID":"c3b6f338-04d7-40c8-baa6-8f17a228a458","Type":"ContainerStarted","Data":"5537f7069b60a13c560d49d54b16ecf99cc41a1342f366e44c39e3093c92b627"} Feb 20 06:51:45 crc kubenswrapper[4492]: I0220 06:51:45.610555 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-28vtj"] Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.072341 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.087220 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7837488e-131f-4866-9417-5c9605902130-memberlist\") pod \"speaker-rnqvv\" (UID: \"7837488e-131f-4866-9417-5c9605902130\") " pod="metallb-system/speaker-rnqvv" Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.109909 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rnqvv" Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.326572 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rnqvv" event={"ID":"7837488e-131f-4866-9417-5c9605902130","Type":"ContainerStarted","Data":"ad5b1101a065664cb8c80f6d9b44d7e1403ad2d2a384a24810ab9986351f9789"} Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.328795 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-28vtj" event={"ID":"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3","Type":"ContainerStarted","Data":"acd7b9e2f74f333d8fd60cfe7d345278d7dfe55bc8dcb4d3dcef9198cfdd1146"} Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.328824 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-28vtj" event={"ID":"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3","Type":"ContainerStarted","Data":"1d27b5d8dc4031110e13a76f523a0c2d65fe4d87c34fa6abd943b11471346d16"} Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.328835 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-28vtj" event={"ID":"6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3","Type":"ContainerStarted","Data":"f2d3e63b13e9758fa112f7d5d699f0307e3452036b7a40de235342cd3d443865"} Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.328963 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:46 crc kubenswrapper[4492]: I0220 06:51:46.353560 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-28vtj" podStartSLOduration=2.353541883 podStartE2EDuration="2.353541883s" podCreationTimestamp="2026-02-20 06:51:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:51:46.348928712 +0000 UTC m=+663.120217690" watchObservedRunningTime="2026-02-20 06:51:46.353541883 +0000 UTC m=+663.124830861" Feb 20 06:51:47 crc kubenswrapper[4492]: I0220 06:51:47.337862 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rnqvv" event={"ID":"7837488e-131f-4866-9417-5c9605902130","Type":"ContainerStarted","Data":"84b82383e31ec1bb08c12db84a9a77d9c2a57cd6e3f2f3653fd0ed495fd1f365"} Feb 20 06:51:47 crc kubenswrapper[4492]: I0220 06:51:47.338162 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rnqvv" event={"ID":"7837488e-131f-4866-9417-5c9605902130","Type":"ContainerStarted","Data":"707e81c3e94e6153d96283f2f859bd3807e822a1df5750c94b04efdd14f74f96"} Feb 20 06:51:47 crc kubenswrapper[4492]: I0220 06:51:47.354701 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-rnqvv" podStartSLOduration=3.354680762 podStartE2EDuration="3.354680762s" podCreationTimestamp="2026-02-20 06:51:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:51:47.351640985 +0000 UTC m=+664.122929964" watchObservedRunningTime="2026-02-20 06:51:47.354680762 +0000 UTC m=+664.125969741" Feb 20 06:51:48 crc kubenswrapper[4492]: I0220 06:51:48.343606 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-rnqvv" Feb 20 06:51:52 crc kubenswrapper[4492]: I0220 06:51:52.381131 4492 generic.go:334] "Generic (PLEG): container finished" podID="1224d7af-027f-4a22-b675-606a9650f716" containerID="0b71bf8eb18accf9e154efb7f4fdf779e3a7357ccf5752adb64b41ea6236e91b" exitCode=0 Feb 20 06:51:52 crc kubenswrapper[4492]: I0220 06:51:52.381248 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerDied","Data":"0b71bf8eb18accf9e154efb7f4fdf779e3a7357ccf5752adb64b41ea6236e91b"} Feb 20 06:51:52 crc kubenswrapper[4492]: I0220 06:51:52.384424 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" event={"ID":"c3b6f338-04d7-40c8-baa6-8f17a228a458","Type":"ContainerStarted","Data":"e383f75ef373a8ed81135e30c0618f3fc76c39512b4f35c4300eab8bab34ad70"} Feb 20 06:51:52 crc kubenswrapper[4492]: I0220 06:51:52.384651 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:51:53 crc kubenswrapper[4492]: I0220 06:51:53.393950 4492 generic.go:334] "Generic (PLEG): container finished" podID="1224d7af-027f-4a22-b675-606a9650f716" containerID="d73d45a7d094a6b2b944951eaf00b77800a1523a4e2a407490511f2721398a80" exitCode=0 Feb 20 06:51:53 crc kubenswrapper[4492]: I0220 06:51:53.394006 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerDied","Data":"d73d45a7d094a6b2b944951eaf00b77800a1523a4e2a407490511f2721398a80"} Feb 20 06:51:53 crc kubenswrapper[4492]: I0220 06:51:53.424753 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" podStartSLOduration=2.032940114 podStartE2EDuration="9.424723292s" podCreationTimestamp="2026-02-20 06:51:44 +0000 UTC" firstStartedPulling="2026-02-20 06:51:44.750861044 +0000 UTC m=+661.522150022" lastFinishedPulling="2026-02-20 06:51:52.142644221 +0000 UTC m=+668.913933200" observedRunningTime="2026-02-20 06:51:52.426536104 +0000 UTC m=+669.197825082" watchObservedRunningTime="2026-02-20 06:51:53.424723292 +0000 UTC m=+670.196012269" Feb 20 06:51:54 crc kubenswrapper[4492]: I0220 06:51:54.405035 4492 generic.go:334] "Generic (PLEG): container finished" podID="1224d7af-027f-4a22-b675-606a9650f716" containerID="30b191cd386c3fa0937cc46bb385608fab5c9d001e4eb3326df502f14e721290" exitCode=0 Feb 20 06:51:54 crc kubenswrapper[4492]: I0220 06:51:54.405102 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerDied","Data":"30b191cd386c3fa0937cc46bb385608fab5c9d001e4eb3326df502f14e721290"} Feb 20 06:51:55 crc kubenswrapper[4492]: I0220 06:51:55.224907 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-28vtj" Feb 20 06:51:55 crc kubenswrapper[4492]: I0220 06:51:55.420935 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerStarted","Data":"f133e3437c391088725d515256fb9c95d8a0f129a8dcf806e82be4ead1fdfdf3"} Feb 20 06:51:55 crc kubenswrapper[4492]: I0220 06:51:55.421000 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerStarted","Data":"0049b814f3f982c41159572be91295cedc3d2eed76fb4d513d9794ee2c293825"} Feb 20 06:51:55 crc kubenswrapper[4492]: I0220 06:51:55.421012 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerStarted","Data":"0944add31886cde0889dfcf44bcd8dcc6be9dbfa3e3920125eb6b99567632ad2"} Feb 20 06:51:56 crc kubenswrapper[4492]: I0220 06:51:56.113878 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-rnqvv" Feb 20 06:51:56 crc kubenswrapper[4492]: I0220 06:51:56.432682 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerStarted","Data":"b061f8b0934ca67c278b7193c0d6acf706c2929f25cbbcd95986a15df014d243"} Feb 20 06:51:56 crc kubenswrapper[4492]: I0220 06:51:56.432739 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerStarted","Data":"e49cf261a5379e397da1f94d3cf7e60417f34cf08eed41589eeab8adc958ce53"} Feb 20 06:51:56 crc kubenswrapper[4492]: I0220 06:51:56.432750 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sps8s" event={"ID":"1224d7af-027f-4a22-b675-606a9650f716","Type":"ContainerStarted","Data":"29c2eabddac6099602c07306f77212eba110ed5fa5482ce3e0e16576a9ee2363"} Feb 20 06:51:56 crc kubenswrapper[4492]: I0220 06:51:56.433083 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.221026 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-sps8s" podStartSLOduration=6.711875278 podStartE2EDuration="14.22100322s" podCreationTimestamp="2026-02-20 06:51:44 +0000 UTC" firstStartedPulling="2026-02-20 06:51:44.656280992 +0000 UTC m=+661.427569970" lastFinishedPulling="2026-02-20 06:51:52.165408934 +0000 UTC m=+668.936697912" observedRunningTime="2026-02-20 06:51:56.456735386 +0000 UTC m=+673.228024365" watchObservedRunningTime="2026-02-20 06:51:58.22100322 +0000 UTC m=+674.992292199" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.228044 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-86gxk"] Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.229248 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-86gxk" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.231705 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.232019 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-86gxk"] Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.232181 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-cskbm" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.232377 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.356385 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nm94\" (UniqueName: \"kubernetes.io/projected/2954167f-4db8-4991-88bc-8a976e02011d-kube-api-access-8nm94\") pod \"openstack-operator-index-86gxk\" (UID: \"2954167f-4db8-4991-88bc-8a976e02011d\") " pod="openstack-operators/openstack-operator-index-86gxk" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.457336 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nm94\" (UniqueName: \"kubernetes.io/projected/2954167f-4db8-4991-88bc-8a976e02011d-kube-api-access-8nm94\") pod \"openstack-operator-index-86gxk\" (UID: \"2954167f-4db8-4991-88bc-8a976e02011d\") " pod="openstack-operators/openstack-operator-index-86gxk" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.475537 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nm94\" (UniqueName: \"kubernetes.io/projected/2954167f-4db8-4991-88bc-8a976e02011d-kube-api-access-8nm94\") pod \"openstack-operator-index-86gxk\" (UID: \"2954167f-4db8-4991-88bc-8a976e02011d\") " pod="openstack-operators/openstack-operator-index-86gxk" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.545831 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-86gxk" Feb 20 06:51:58 crc kubenswrapper[4492]: I0220 06:51:58.756731 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-86gxk"] Feb 20 06:51:59 crc kubenswrapper[4492]: I0220 06:51:59.452829 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-86gxk" event={"ID":"2954167f-4db8-4991-88bc-8a976e02011d","Type":"ContainerStarted","Data":"e96deaa95225737d0d67dbc538c4cdc202ce97758115ce9f6cd79d1d01061229"} Feb 20 06:51:59 crc kubenswrapper[4492]: I0220 06:51:59.524721 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-sps8s" Feb 20 06:51:59 crc kubenswrapper[4492]: I0220 06:51:59.622142 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-sps8s" Feb 20 06:52:00 crc kubenswrapper[4492]: I0220 06:52:00.461165 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-86gxk" event={"ID":"2954167f-4db8-4991-88bc-8a976e02011d","Type":"ContainerStarted","Data":"4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682"} Feb 20 06:52:00 crc kubenswrapper[4492]: I0220 06:52:00.499927 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-86gxk" podStartSLOduration=1.134016908 podStartE2EDuration="2.499902873s" podCreationTimestamp="2026-02-20 06:51:58 +0000 UTC" firstStartedPulling="2026-02-20 06:51:58.772772081 +0000 UTC m=+675.544061059" lastFinishedPulling="2026-02-20 06:52:00.138658046 +0000 UTC m=+676.909947024" observedRunningTime="2026-02-20 06:52:00.497609221 +0000 UTC m=+677.268898189" watchObservedRunningTime="2026-02-20 06:52:00.499902873 +0000 UTC m=+677.271191851" Feb 20 06:52:01 crc kubenswrapper[4492]: I0220 06:52:01.389145 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-86gxk"] Feb 20 06:52:01 crc kubenswrapper[4492]: I0220 06:52:01.997838 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4hv4z"] Feb 20 06:52:01 crc kubenswrapper[4492]: I0220 06:52:01.999444 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.010971 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfcwx\" (UniqueName: \"kubernetes.io/projected/7fd5d9fb-a07c-4715-b281-d181fd8f5543-kube-api-access-vfcwx\") pod \"openstack-operator-index-4hv4z\" (UID: \"7fd5d9fb-a07c-4715-b281-d181fd8f5543\") " pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.020510 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4hv4z"] Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.112258 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfcwx\" (UniqueName: \"kubernetes.io/projected/7fd5d9fb-a07c-4715-b281-d181fd8f5543-kube-api-access-vfcwx\") pod \"openstack-operator-index-4hv4z\" (UID: \"7fd5d9fb-a07c-4715-b281-d181fd8f5543\") " pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.131066 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfcwx\" (UniqueName: \"kubernetes.io/projected/7fd5d9fb-a07c-4715-b281-d181fd8f5543-kube-api-access-vfcwx\") pod \"openstack-operator-index-4hv4z\" (UID: \"7fd5d9fb-a07c-4715-b281-d181fd8f5543\") " pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.314316 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.474908 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-86gxk" podUID="2954167f-4db8-4991-88bc-8a976e02011d" containerName="registry-server" containerID="cri-o://4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682" gracePeriod=2 Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.697858 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4hv4z"] Feb 20 06:52:02 crc kubenswrapper[4492]: W0220 06:52:02.716529 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fd5d9fb_a07c_4715_b281_d181fd8f5543.slice/crio-420c419094a78505eed1d17b534264145645f0988a5e745671ecfdd68228c0ba WatchSource:0}: Error finding container 420c419094a78505eed1d17b534264145645f0988a5e745671ecfdd68228c0ba: Status 404 returned error can't find the container with id 420c419094a78505eed1d17b534264145645f0988a5e745671ecfdd68228c0ba Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.789442 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-86gxk" Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.923497 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nm94\" (UniqueName: \"kubernetes.io/projected/2954167f-4db8-4991-88bc-8a976e02011d-kube-api-access-8nm94\") pod \"2954167f-4db8-4991-88bc-8a976e02011d\" (UID: \"2954167f-4db8-4991-88bc-8a976e02011d\") " Feb 20 06:52:02 crc kubenswrapper[4492]: I0220 06:52:02.928486 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2954167f-4db8-4991-88bc-8a976e02011d-kube-api-access-8nm94" (OuterVolumeSpecName: "kube-api-access-8nm94") pod "2954167f-4db8-4991-88bc-8a976e02011d" (UID: "2954167f-4db8-4991-88bc-8a976e02011d"). InnerVolumeSpecName "kube-api-access-8nm94". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.024642 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nm94\" (UniqueName: \"kubernetes.io/projected/2954167f-4db8-4991-88bc-8a976e02011d-kube-api-access-8nm94\") on node \"crc\" DevicePath \"\"" Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.484811 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4hv4z" event={"ID":"7fd5d9fb-a07c-4715-b281-d181fd8f5543","Type":"ContainerStarted","Data":"32ef55d6805fa1b740df124f783938b9ca71f54355a809463df02318d5c49189"} Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.484877 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4hv4z" event={"ID":"7fd5d9fb-a07c-4715-b281-d181fd8f5543","Type":"ContainerStarted","Data":"420c419094a78505eed1d17b534264145645f0988a5e745671ecfdd68228c0ba"} Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.486347 4492 generic.go:334] "Generic (PLEG): container finished" podID="2954167f-4db8-4991-88bc-8a976e02011d" containerID="4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682" exitCode=0 Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.486397 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-86gxk" Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.486413 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-86gxk" event={"ID":"2954167f-4db8-4991-88bc-8a976e02011d","Type":"ContainerDied","Data":"4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682"} Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.486449 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-86gxk" event={"ID":"2954167f-4db8-4991-88bc-8a976e02011d","Type":"ContainerDied","Data":"e96deaa95225737d0d67dbc538c4cdc202ce97758115ce9f6cd79d1d01061229"} Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.486468 4492 scope.go:117] "RemoveContainer" containerID="4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682" Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.507371 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4hv4z" podStartSLOduration=2.018918929 podStartE2EDuration="2.50734803s" podCreationTimestamp="2026-02-20 06:52:01 +0000 UTC" firstStartedPulling="2026-02-20 06:52:02.720849845 +0000 UTC m=+679.492138823" lastFinishedPulling="2026-02-20 06:52:03.209278946 +0000 UTC m=+679.980567924" observedRunningTime="2026-02-20 06:52:03.507094343 +0000 UTC m=+680.278383321" watchObservedRunningTime="2026-02-20 06:52:03.50734803 +0000 UTC m=+680.278637009" Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.514000 4492 scope.go:117] "RemoveContainer" containerID="4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682" Feb 20 06:52:03 crc kubenswrapper[4492]: E0220 06:52:03.514766 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682\": container with ID starting with 4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682 not found: ID does not exist" containerID="4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682" Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.514825 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682"} err="failed to get container status \"4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682\": rpc error: code = NotFound desc = could not find container \"4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682\": container with ID starting with 4e33bd96fb6c63e3c5730ef89271892a977106573444eb7d6ec0bb84af63d682 not found: ID does not exist" Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.522639 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-86gxk"] Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.526449 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-86gxk"] Feb 20 06:52:03 crc kubenswrapper[4492]: I0220 06:52:03.564721 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2954167f-4db8-4991-88bc-8a976e02011d" path="/var/lib/kubelet/pods/2954167f-4db8-4991-88bc-8a976e02011d/volumes" Feb 20 06:52:04 crc kubenswrapper[4492]: I0220 06:52:04.541618 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-lmpxr" Feb 20 06:52:09 crc kubenswrapper[4492]: I0220 06:52:09.311332 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:52:09 crc kubenswrapper[4492]: I0220 06:52:09.312060 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:52:12 crc kubenswrapper[4492]: I0220 06:52:12.314442 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:12 crc kubenswrapper[4492]: I0220 06:52:12.315111 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:12 crc kubenswrapper[4492]: I0220 06:52:12.343225 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:12 crc kubenswrapper[4492]: I0220 06:52:12.585673 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-4hv4z" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.635217 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6"] Feb 20 06:52:13 crc kubenswrapper[4492]: E0220 06:52:13.636089 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2954167f-4db8-4991-88bc-8a976e02011d" containerName="registry-server" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.636157 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2954167f-4db8-4991-88bc-8a976e02011d" containerName="registry-server" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.636420 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2954167f-4db8-4991-88bc-8a976e02011d" containerName="registry-server" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.637388 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.640431 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-zbs8p" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.648369 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6"] Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.753978 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-bundle\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.754110 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sxx9\" (UniqueName: \"kubernetes.io/projected/dc079aa5-a7a5-4043-951b-65da547fa1e1-kube-api-access-5sxx9\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.754286 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-util\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.856029 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-bundle\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.856090 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sxx9\" (UniqueName: \"kubernetes.io/projected/dc079aa5-a7a5-4043-951b-65da547fa1e1-kube-api-access-5sxx9\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.856168 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-util\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.856575 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-bundle\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.856671 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-util\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.874038 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sxx9\" (UniqueName: \"kubernetes.io/projected/dc079aa5-a7a5-4043-951b-65da547fa1e1-kube-api-access-5sxx9\") pod \"8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:13 crc kubenswrapper[4492]: I0220 06:52:13.969752 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:14 crc kubenswrapper[4492]: I0220 06:52:14.356357 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6"] Feb 20 06:52:14 crc kubenswrapper[4492]: W0220 06:52:14.361040 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc079aa5_a7a5_4043_951b_65da547fa1e1.slice/crio-1dbc824eff2e7b6555d40bcf62150556757a1f29a7fa504c1c23383a01889300 WatchSource:0}: Error finding container 1dbc824eff2e7b6555d40bcf62150556757a1f29a7fa504c1c23383a01889300: Status 404 returned error can't find the container with id 1dbc824eff2e7b6555d40bcf62150556757a1f29a7fa504c1c23383a01889300 Feb 20 06:52:14 crc kubenswrapper[4492]: I0220 06:52:14.528138 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-sps8s" Feb 20 06:52:14 crc kubenswrapper[4492]: I0220 06:52:14.571080 4492 generic.go:334] "Generic (PLEG): container finished" podID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerID="daf06995ad8b7f1ef02afabce6fedd938d2544568f60c17962a9a068391ce8e4" exitCode=0 Feb 20 06:52:14 crc kubenswrapper[4492]: I0220 06:52:14.571123 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" event={"ID":"dc079aa5-a7a5-4043-951b-65da547fa1e1","Type":"ContainerDied","Data":"daf06995ad8b7f1ef02afabce6fedd938d2544568f60c17962a9a068391ce8e4"} Feb 20 06:52:14 crc kubenswrapper[4492]: I0220 06:52:14.571143 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" event={"ID":"dc079aa5-a7a5-4043-951b-65da547fa1e1","Type":"ContainerStarted","Data":"1dbc824eff2e7b6555d40bcf62150556757a1f29a7fa504c1c23383a01889300"} Feb 20 06:52:16 crc kubenswrapper[4492]: I0220 06:52:16.586138 4492 generic.go:334] "Generic (PLEG): container finished" podID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerID="b9f401eff701a33a40cb0d4b959e36567173e913bc5522626d99081ac0ad0baa" exitCode=0 Feb 20 06:52:16 crc kubenswrapper[4492]: I0220 06:52:16.586220 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" event={"ID":"dc079aa5-a7a5-4043-951b-65da547fa1e1","Type":"ContainerDied","Data":"b9f401eff701a33a40cb0d4b959e36567173e913bc5522626d99081ac0ad0baa"} Feb 20 06:52:17 crc kubenswrapper[4492]: I0220 06:52:17.608990 4492 generic.go:334] "Generic (PLEG): container finished" podID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerID="83f832a5eeab288e3ddc8036765c394eb4844c132f5594238c2829069f26e334" exitCode=0 Feb 20 06:52:17 crc kubenswrapper[4492]: I0220 06:52:17.609080 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" event={"ID":"dc079aa5-a7a5-4043-951b-65da547fa1e1","Type":"ContainerDied","Data":"83f832a5eeab288e3ddc8036765c394eb4844c132f5594238c2829069f26e334"} Feb 20 06:52:18 crc kubenswrapper[4492]: I0220 06:52:18.832148 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.027272 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sxx9\" (UniqueName: \"kubernetes.io/projected/dc079aa5-a7a5-4043-951b-65da547fa1e1-kube-api-access-5sxx9\") pod \"dc079aa5-a7a5-4043-951b-65da547fa1e1\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.027511 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-util\") pod \"dc079aa5-a7a5-4043-951b-65da547fa1e1\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.027560 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-bundle\") pod \"dc079aa5-a7a5-4043-951b-65da547fa1e1\" (UID: \"dc079aa5-a7a5-4043-951b-65da547fa1e1\") " Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.029271 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-bundle" (OuterVolumeSpecName: "bundle") pod "dc079aa5-a7a5-4043-951b-65da547fa1e1" (UID: "dc079aa5-a7a5-4043-951b-65da547fa1e1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.033762 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc079aa5-a7a5-4043-951b-65da547fa1e1-kube-api-access-5sxx9" (OuterVolumeSpecName: "kube-api-access-5sxx9") pod "dc079aa5-a7a5-4043-951b-65da547fa1e1" (UID: "dc079aa5-a7a5-4043-951b-65da547fa1e1"). InnerVolumeSpecName "kube-api-access-5sxx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.038895 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-util" (OuterVolumeSpecName: "util") pod "dc079aa5-a7a5-4043-951b-65da547fa1e1" (UID: "dc079aa5-a7a5-4043-951b-65da547fa1e1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.129777 4492 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-util\") on node \"crc\" DevicePath \"\"" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.129804 4492 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dc079aa5-a7a5-4043-951b-65da547fa1e1-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.129823 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sxx9\" (UniqueName: \"kubernetes.io/projected/dc079aa5-a7a5-4043-951b-65da547fa1e1-kube-api-access-5sxx9\") on node \"crc\" DevicePath \"\"" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.622546 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" event={"ID":"dc079aa5-a7a5-4043-951b-65da547fa1e1","Type":"ContainerDied","Data":"1dbc824eff2e7b6555d40bcf62150556757a1f29a7fa504c1c23383a01889300"} Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.622599 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dbc824eff2e7b6555d40bcf62150556757a1f29a7fa504c1c23383a01889300" Feb 20 06:52:19 crc kubenswrapper[4492]: I0220 06:52:19.622621 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.864202 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh"] Feb 20 06:52:21 crc kubenswrapper[4492]: E0220 06:52:21.865035 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerName="pull" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.865058 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerName="pull" Feb 20 06:52:21 crc kubenswrapper[4492]: E0220 06:52:21.865079 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerName="extract" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.865086 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerName="extract" Feb 20 06:52:21 crc kubenswrapper[4492]: E0220 06:52:21.865106 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerName="util" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.865115 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerName="util" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.865300 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc079aa5-a7a5-4043-951b-65da547fa1e1" containerName="extract" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.865918 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.869146 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-7mrtw" Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.901079 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh"] Feb 20 06:52:21 crc kubenswrapper[4492]: I0220 06:52:21.962729 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlxj4\" (UniqueName: \"kubernetes.io/projected/c7b6f955-806d-4f15-bc40-8fccbdeefa08-kube-api-access-jlxj4\") pod \"openstack-operator-controller-init-6679bf9b57-hd7zh\" (UID: \"c7b6f955-806d-4f15-bc40-8fccbdeefa08\") " pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" Feb 20 06:52:22 crc kubenswrapper[4492]: I0220 06:52:22.064778 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlxj4\" (UniqueName: \"kubernetes.io/projected/c7b6f955-806d-4f15-bc40-8fccbdeefa08-kube-api-access-jlxj4\") pod \"openstack-operator-controller-init-6679bf9b57-hd7zh\" (UID: \"c7b6f955-806d-4f15-bc40-8fccbdeefa08\") " pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" Feb 20 06:52:22 crc kubenswrapper[4492]: I0220 06:52:22.084463 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlxj4\" (UniqueName: \"kubernetes.io/projected/c7b6f955-806d-4f15-bc40-8fccbdeefa08-kube-api-access-jlxj4\") pod \"openstack-operator-controller-init-6679bf9b57-hd7zh\" (UID: \"c7b6f955-806d-4f15-bc40-8fccbdeefa08\") " pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" Feb 20 06:52:22 crc kubenswrapper[4492]: I0220 06:52:22.182296 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" Feb 20 06:52:22 crc kubenswrapper[4492]: I0220 06:52:22.589153 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh"] Feb 20 06:52:22 crc kubenswrapper[4492]: I0220 06:52:22.643071 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" event={"ID":"c7b6f955-806d-4f15-bc40-8fccbdeefa08","Type":"ContainerStarted","Data":"7f0af5c45a304a4da14224cc35ccdc15d51097ed79671057e2919e0464bc3fa8"} Feb 20 06:52:28 crc kubenswrapper[4492]: I0220 06:52:28.694418 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" event={"ID":"c7b6f955-806d-4f15-bc40-8fccbdeefa08","Type":"ContainerStarted","Data":"aba34bc198f122416474bf79288cdb538b5b26969ca5129c2f76600fe3748c48"} Feb 20 06:52:28 crc kubenswrapper[4492]: I0220 06:52:28.695062 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" Feb 20 06:52:39 crc kubenswrapper[4492]: I0220 06:52:39.311857 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:52:39 crc kubenswrapper[4492]: I0220 06:52:39.312653 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:52:42 crc kubenswrapper[4492]: I0220 06:52:42.186338 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" Feb 20 06:52:42 crc kubenswrapper[4492]: I0220 06:52:42.215684 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6679bf9b57-hd7zh" podStartSLOduration=15.715281821 podStartE2EDuration="21.215662275s" podCreationTimestamp="2026-02-20 06:52:21 +0000 UTC" firstStartedPulling="2026-02-20 06:52:22.614543096 +0000 UTC m=+699.385832074" lastFinishedPulling="2026-02-20 06:52:28.114923551 +0000 UTC m=+704.886212528" observedRunningTime="2026-02-20 06:52:28.718706673 +0000 UTC m=+705.489995652" watchObservedRunningTime="2026-02-20 06:52:42.215662275 +0000 UTC m=+718.986951253" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.480575 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.481795 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.483697 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hstlw" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.494933 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.506190 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.507281 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.516061 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-f6dqh" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.527660 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.528543 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.530018 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.530341 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-4cdcl" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.534224 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.544592 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.545208 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.549210 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mdnw2" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.607640 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.614589 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.614703 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.629174 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-c9zm8" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.629904 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.664642 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.665731 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.672150 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcvzd\" (UniqueName: \"kubernetes.io/projected/d1f3b23e-c05b-42f0-af00-5af28a768c0e-kube-api-access-zcvzd\") pod \"barbican-operator-controller-manager-868647ff47-qrmfx\" (UID: \"d1f3b23e-c05b-42f0-af00-5af28a768c0e\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.672202 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lgf8\" (UniqueName: \"kubernetes.io/projected/16dae257-55aa-40f7-9308-182cd7ff886b-kube-api-access-2lgf8\") pod \"designate-operator-controller-manager-6d8bf5c495-h27fh\" (UID: \"16dae257-55aa-40f7-9308-182cd7ff886b\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.672231 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c2g7\" (UniqueName: \"kubernetes.io/projected/1f1dd67b-1f84-49a3-8ffa-bba932485c90-kube-api-access-4c2g7\") pod \"cinder-operator-controller-manager-5d946d989d-vqxgx\" (UID: \"1f1dd67b-1f84-49a3-8ffa-bba932485c90\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.672346 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5xfr\" (UniqueName: \"kubernetes.io/projected/dd787991-93c0-4baf-8a53-c2ec22c09ede-kube-api-access-g5xfr\") pod \"glance-operator-controller-manager-77987464f4-f5lt2\" (UID: \"dd787991-93c0-4baf-8a53-c2ec22c09ede\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.674359 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.675263 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.678349 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-422nx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.684228 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-szmrb"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.685279 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.685915 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-rkzzv" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.694009 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.694308 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.699382 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-8rsw7" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.710571 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.715957 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.716646 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-szmrb"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.716726 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.723395 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.723617 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-662fl" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.724391 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.726348 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.726417 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8x7kd" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.727242 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.728405 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.728998 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.735410 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.736222 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.736322 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-6jbkl" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.736566 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qnn2s" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.740099 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.748572 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.755833 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.763933 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.764031 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.778130 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-rq4rj" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.779394 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pkq2\" (UniqueName: \"kubernetes.io/projected/80579e11-ccf9-4377-a013-345a55ac33ab-kube-api-access-6pkq2\") pod \"heat-operator-controller-manager-69f49c598c-5kqdb\" (UID: \"80579e11-ccf9-4377-a013-345a55ac33ab\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.779549 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lgf8\" (UniqueName: \"kubernetes.io/projected/16dae257-55aa-40f7-9308-182cd7ff886b-kube-api-access-2lgf8\") pod \"designate-operator-controller-manager-6d8bf5c495-h27fh\" (UID: \"16dae257-55aa-40f7-9308-182cd7ff886b\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.779642 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c2g7\" (UniqueName: \"kubernetes.io/projected/1f1dd67b-1f84-49a3-8ffa-bba932485c90-kube-api-access-4c2g7\") pod \"cinder-operator-controller-manager-5d946d989d-vqxgx\" (UID: \"1f1dd67b-1f84-49a3-8ffa-bba932485c90\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.779757 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g2jf\" (UniqueName: \"kubernetes.io/projected/9c26743a-8dbb-4bdc-a2a8-430efd8070c6-kube-api-access-9g2jf\") pod \"horizon-operator-controller-manager-5b9b8895d5-jm5rz\" (UID: \"9c26743a-8dbb-4bdc-a2a8-430efd8070c6\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.779870 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5xfr\" (UniqueName: \"kubernetes.io/projected/dd787991-93c0-4baf-8a53-c2ec22c09ede-kube-api-access-g5xfr\") pod \"glance-operator-controller-manager-77987464f4-f5lt2\" (UID: \"dd787991-93c0-4baf-8a53-c2ec22c09ede\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.779972 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcvzd\" (UniqueName: \"kubernetes.io/projected/d1f3b23e-c05b-42f0-af00-5af28a768c0e-kube-api-access-zcvzd\") pod \"barbican-operator-controller-manager-868647ff47-qrmfx\" (UID: \"d1f3b23e-c05b-42f0-af00-5af28a768c0e\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.806482 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.807240 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.808748 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcvzd\" (UniqueName: \"kubernetes.io/projected/d1f3b23e-c05b-42f0-af00-5af28a768c0e-kube-api-access-zcvzd\") pod \"barbican-operator-controller-manager-868647ff47-qrmfx\" (UID: \"d1f3b23e-c05b-42f0-af00-5af28a768c0e\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.816131 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.816978 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.821991 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-2hbvq" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.827158 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lgf8\" (UniqueName: \"kubernetes.io/projected/16dae257-55aa-40f7-9308-182cd7ff886b-kube-api-access-2lgf8\") pod \"designate-operator-controller-manager-6d8bf5c495-h27fh\" (UID: \"16dae257-55aa-40f7-9308-182cd7ff886b\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.827397 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c2g7\" (UniqueName: \"kubernetes.io/projected/1f1dd67b-1f84-49a3-8ffa-bba932485c90-kube-api-access-4c2g7\") pod \"cinder-operator-controller-manager-5d946d989d-vqxgx\" (UID: \"1f1dd67b-1f84-49a3-8ffa-bba932485c90\") " pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.827804 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.839841 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-24jdc" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.840232 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5xfr\" (UniqueName: \"kubernetes.io/projected/dd787991-93c0-4baf-8a53-c2ec22c09ede-kube-api-access-g5xfr\") pod \"glance-operator-controller-manager-77987464f4-f5lt2\" (UID: \"dd787991-93c0-4baf-8a53-c2ec22c09ede\") " pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.844998 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.848500 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.865414 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.886963 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889424 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58r6t\" (UniqueName: \"kubernetes.io/projected/86409721-594f-4c20-9f11-2f89572c3036-kube-api-access-58r6t\") pod \"ironic-operator-controller-manager-554564d7fc-s7bxs\" (UID: \"86409721-594f-4c20-9f11-2f89572c3036\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889468 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm6zm\" (UniqueName: \"kubernetes.io/projected/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-kube-api-access-wm6zm\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889510 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889526 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbkkl\" (UniqueName: \"kubernetes.io/projected/2918bb03-8106-4586-90bc-2618b3e78239-kube-api-access-bbkkl\") pod \"keystone-operator-controller-manager-b4d948c87-t8cd6\" (UID: \"2918bb03-8106-4586-90bc-2618b3e78239\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889551 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pkq2\" (UniqueName: \"kubernetes.io/projected/80579e11-ccf9-4377-a013-345a55ac33ab-kube-api-access-6pkq2\") pod \"heat-operator-controller-manager-69f49c598c-5kqdb\" (UID: \"80579e11-ccf9-4377-a013-345a55ac33ab\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889570 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889586 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrnnq\" (UniqueName: \"kubernetes.io/projected/b1cfd35f-eae6-453c-8687-6bf96945edf2-kube-api-access-zrnnq\") pod \"mariadb-operator-controller-manager-6994f66f48-dnt2c\" (UID: \"b1cfd35f-eae6-453c-8687-6bf96945edf2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889609 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l965\" (UniqueName: \"kubernetes.io/projected/7e6bf429-fb74-4a04-ac01-b4f60d4c00c2-kube-api-access-8l965\") pod \"manila-operator-controller-manager-54f6768c69-dntvf\" (UID: \"7e6bf429-fb74-4a04-ac01-b4f60d4c00c2\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889624 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6dlc\" (UniqueName: \"kubernetes.io/projected/cafb9e27-7f46-4d0d-a27f-7f60d11d5b79-kube-api-access-c6dlc\") pod \"octavia-operator-controller-manager-69f8888797-f6zqc\" (UID: \"cafb9e27-7f46-4d0d-a27f-7f60d11d5b79\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889639 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5g76\" (UniqueName: \"kubernetes.io/projected/883be441-1faf-4b78-9c22-16c19986902a-kube-api-access-b5g76\") pod \"nova-operator-controller-manager-567668f5cf-2zzhs\" (UID: \"883be441-1faf-4b78-9c22-16c19986902a\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889660 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4bgg\" (UniqueName: \"kubernetes.io/projected/ea530a5e-2861-4559-b80a-ef8589d9788f-kube-api-access-g4bgg\") pod \"neutron-operator-controller-manager-64ddbf8bb-vckjp\" (UID: \"ea530a5e-2861-4559-b80a-ef8589d9788f\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889696 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g2jf\" (UniqueName: \"kubernetes.io/projected/9c26743a-8dbb-4bdc-a2a8-430efd8070c6-kube-api-access-9g2jf\") pod \"horizon-operator-controller-manager-5b9b8895d5-jm5rz\" (UID: \"9c26743a-8dbb-4bdc-a2a8-430efd8070c6\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.889717 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5h8j\" (UniqueName: \"kubernetes.io/projected/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-kube-api-access-w5h8j\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.901446 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.902800 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.919975 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g2jf\" (UniqueName: \"kubernetes.io/projected/9c26743a-8dbb-4bdc-a2a8-430efd8070c6-kube-api-access-9g2jf\") pod \"horizon-operator-controller-manager-5b9b8895d5-jm5rz\" (UID: \"9c26743a-8dbb-4bdc-a2a8-430efd8070c6\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.920731 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-z96cf" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.921287 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.944156 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt"] Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.945818 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.947952 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-hnfjx" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.954448 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pkq2\" (UniqueName: \"kubernetes.io/projected/80579e11-ccf9-4377-a013-345a55ac33ab-kube-api-access-6pkq2\") pod \"heat-operator-controller-manager-69f49c598c-5kqdb\" (UID: \"80579e11-ccf9-4377-a013-345a55ac33ab\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" Feb 20 06:53:01 crc kubenswrapper[4492]: I0220 06:53:01.966108 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.002796 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003387 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58r6t\" (UniqueName: \"kubernetes.io/projected/86409721-594f-4c20-9f11-2f89572c3036-kube-api-access-58r6t\") pod \"ironic-operator-controller-manager-554564d7fc-s7bxs\" (UID: \"86409721-594f-4c20-9f11-2f89572c3036\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003496 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm6zm\" (UniqueName: \"kubernetes.io/projected/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-kube-api-access-wm6zm\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003549 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003571 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbkkl\" (UniqueName: \"kubernetes.io/projected/2918bb03-8106-4586-90bc-2618b3e78239-kube-api-access-bbkkl\") pod \"keystone-operator-controller-manager-b4d948c87-t8cd6\" (UID: \"2918bb03-8106-4586-90bc-2618b3e78239\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003601 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003624 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrnnq\" (UniqueName: \"kubernetes.io/projected/b1cfd35f-eae6-453c-8687-6bf96945edf2-kube-api-access-zrnnq\") pod \"mariadb-operator-controller-manager-6994f66f48-dnt2c\" (UID: \"b1cfd35f-eae6-453c-8687-6bf96945edf2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003655 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l965\" (UniqueName: \"kubernetes.io/projected/7e6bf429-fb74-4a04-ac01-b4f60d4c00c2-kube-api-access-8l965\") pod \"manila-operator-controller-manager-54f6768c69-dntvf\" (UID: \"7e6bf429-fb74-4a04-ac01-b4f60d4c00c2\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003671 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6dlc\" (UniqueName: \"kubernetes.io/projected/cafb9e27-7f46-4d0d-a27f-7f60d11d5b79-kube-api-access-c6dlc\") pod \"octavia-operator-controller-manager-69f8888797-f6zqc\" (UID: \"cafb9e27-7f46-4d0d-a27f-7f60d11d5b79\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003691 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5g76\" (UniqueName: \"kubernetes.io/projected/883be441-1faf-4b78-9c22-16c19986902a-kube-api-access-b5g76\") pod \"nova-operator-controller-manager-567668f5cf-2zzhs\" (UID: \"883be441-1faf-4b78-9c22-16c19986902a\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003726 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4bgg\" (UniqueName: \"kubernetes.io/projected/ea530a5e-2861-4559-b80a-ef8589d9788f-kube-api-access-g4bgg\") pod \"neutron-operator-controller-manager-64ddbf8bb-vckjp\" (UID: \"ea530a5e-2861-4559-b80a-ef8589d9788f\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.003764 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5h8j\" (UniqueName: \"kubernetes.io/projected/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-kube-api-access-w5h8j\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.003845 4492 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.003887 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert podName:bab88145-f203-4b8c-b9d3-a03ee1af5ac7 nodeName:}" failed. No retries permitted until 2026-02-20 06:53:02.503872527 +0000 UTC m=+739.275161494 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert") pod "infra-operator-controller-manager-79d975b745-szmrb" (UID: "bab88145-f203-4b8c-b9d3-a03ee1af5ac7") : secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.004214 4492 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.004290 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert podName:41543a2e-55d9-465c-bb77-f2d7b04b7e2b nodeName:}" failed. No retries permitted until 2026-02-20 06:53:02.504283541 +0000 UTC m=+739.275572520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" (UID: "41543a2e-55d9-465c-bb77-f2d7b04b7e2b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.014125 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.026986 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58r6t\" (UniqueName: \"kubernetes.io/projected/86409721-594f-4c20-9f11-2f89572c3036-kube-api-access-58r6t\") pod \"ironic-operator-controller-manager-554564d7fc-s7bxs\" (UID: \"86409721-594f-4c20-9f11-2f89572c3036\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.027648 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-77tpg"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.028678 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.037119 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-czrmd" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.043574 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6dlc\" (UniqueName: \"kubernetes.io/projected/cafb9e27-7f46-4d0d-a27f-7f60d11d5b79-kube-api-access-c6dlc\") pod \"octavia-operator-controller-manager-69f8888797-f6zqc\" (UID: \"cafb9e27-7f46-4d0d-a27f-7f60d11d5b79\") " pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.044551 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.046146 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4bgg\" (UniqueName: \"kubernetes.io/projected/ea530a5e-2861-4559-b80a-ef8589d9788f-kube-api-access-g4bgg\") pod \"neutron-operator-controller-manager-64ddbf8bb-vckjp\" (UID: \"ea530a5e-2861-4559-b80a-ef8589d9788f\") " pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.046643 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l965\" (UniqueName: \"kubernetes.io/projected/7e6bf429-fb74-4a04-ac01-b4f60d4c00c2-kube-api-access-8l965\") pod \"manila-operator-controller-manager-54f6768c69-dntvf\" (UID: \"7e6bf429-fb74-4a04-ac01-b4f60d4c00c2\") " pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.046908 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.047301 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrnnq\" (UniqueName: \"kubernetes.io/projected/b1cfd35f-eae6-453c-8687-6bf96945edf2-kube-api-access-zrnnq\") pod \"mariadb-operator-controller-manager-6994f66f48-dnt2c\" (UID: \"b1cfd35f-eae6-453c-8687-6bf96945edf2\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.048808 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-m6lrb" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.052762 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm6zm\" (UniqueName: \"kubernetes.io/projected/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-kube-api-access-wm6zm\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.055960 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.059971 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5g76\" (UniqueName: \"kubernetes.io/projected/883be441-1faf-4b78-9c22-16c19986902a-kube-api-access-b5g76\") pod \"nova-operator-controller-manager-567668f5cf-2zzhs\" (UID: \"883be441-1faf-4b78-9c22-16c19986902a\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.060050 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5h8j\" (UniqueName: \"kubernetes.io/projected/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-kube-api-access-w5h8j\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.070739 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.071289 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.102237 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.103040 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.103104 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.108594 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhdtl\" (UniqueName: \"kubernetes.io/projected/10788676-0981-46c0-8a6a-eda4f457e117-kube-api-access-fhdtl\") pod \"placement-operator-controller-manager-8497b45c89-6hbwt\" (UID: \"10788676-0981-46c0-8a6a-eda4f457e117\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.108716 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dt74\" (UniqueName: \"kubernetes.io/projected/91e88c5d-7cb2-4ebd-9a96-d3620ea38c97-kube-api-access-9dt74\") pod \"ovn-operator-controller-manager-d44cf6b75-t5t76\" (UID: \"91e88c5d-7cb2-4ebd-9a96-d3620ea38c97\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.109755 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbkkl\" (UniqueName: \"kubernetes.io/projected/2918bb03-8106-4586-90bc-2618b3e78239-kube-api-access-bbkkl\") pod \"keystone-operator-controller-manager-b4d948c87-t8cd6\" (UID: \"2918bb03-8106-4586-90bc-2618b3e78239\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.164705 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-77tpg"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.182253 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.206383 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-p5tgj"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.210216 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8pb8\" (UniqueName: \"kubernetes.io/projected/e2b03896-ea8a-469a-8a43-4f04f2108071-kube-api-access-s8pb8\") pod \"telemetry-operator-controller-manager-7f45b4ff68-t4h7v\" (UID: \"e2b03896-ea8a-469a-8a43-4f04f2108071\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.226998 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhdtl\" (UniqueName: \"kubernetes.io/projected/10788676-0981-46c0-8a6a-eda4f457e117-kube-api-access-fhdtl\") pod \"placement-operator-controller-manager-8497b45c89-6hbwt\" (UID: \"10788676-0981-46c0-8a6a-eda4f457e117\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.227079 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjx7f\" (UniqueName: \"kubernetes.io/projected/ba9af5af-1c2b-421c-a59a-4f4a203a8e3e-kube-api-access-jjx7f\") pod \"swift-operator-controller-manager-68f46476f-77tpg\" (UID: \"ba9af5af-1c2b-421c-a59a-4f4a203a8e3e\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.227265 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dt74\" (UniqueName: \"kubernetes.io/projected/91e88c5d-7cb2-4ebd-9a96-d3620ea38c97-kube-api-access-9dt74\") pod \"ovn-operator-controller-manager-d44cf6b75-t5t76\" (UID: \"91e88c5d-7cb2-4ebd-9a96-d3620ea38c97\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.231223 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.241734 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.246771 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-rzbjz" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.262616 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dt74\" (UniqueName: \"kubernetes.io/projected/91e88c5d-7cb2-4ebd-9a96-d3620ea38c97-kube-api-access-9dt74\") pod \"ovn-operator-controller-manager-d44cf6b75-t5t76\" (UID: \"91e88c5d-7cb2-4ebd-9a96-d3620ea38c97\") " pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.264539 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhdtl\" (UniqueName: \"kubernetes.io/projected/10788676-0981-46c0-8a6a-eda4f457e117-kube-api-access-fhdtl\") pod \"placement-operator-controller-manager-8497b45c89-6hbwt\" (UID: \"10788676-0981-46c0-8a6a-eda4f457e117\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.267070 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.276746 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-p5tgj"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.304820 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.316223 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.330276 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8pb8\" (UniqueName: \"kubernetes.io/projected/e2b03896-ea8a-469a-8a43-4f04f2108071-kube-api-access-s8pb8\") pod \"telemetry-operator-controller-manager-7f45b4ff68-t4h7v\" (UID: \"e2b03896-ea8a-469a-8a43-4f04f2108071\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.330324 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjx7f\" (UniqueName: \"kubernetes.io/projected/ba9af5af-1c2b-421c-a59a-4f4a203a8e3e-kube-api-access-jjx7f\") pod \"swift-operator-controller-manager-68f46476f-77tpg\" (UID: \"ba9af5af-1c2b-421c-a59a-4f4a203a8e3e\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.342614 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.343819 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.344327 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.350899 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mmwd2" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.356078 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.357120 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjx7f\" (UniqueName: \"kubernetes.io/projected/ba9af5af-1c2b-421c-a59a-4f4a203a8e3e-kube-api-access-jjx7f\") pod \"swift-operator-controller-manager-68f46476f-77tpg\" (UID: \"ba9af5af-1c2b-421c-a59a-4f4a203a8e3e\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.358173 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8pb8\" (UniqueName: \"kubernetes.io/projected/e2b03896-ea8a-469a-8a43-4f04f2108071-kube-api-access-s8pb8\") pod \"telemetry-operator-controller-manager-7f45b4ff68-t4h7v\" (UID: \"e2b03896-ea8a-469a-8a43-4f04f2108071\") " pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.363763 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.364293 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.364844 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.371906 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.372078 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-lzpf2" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.372208 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.382205 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.408374 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.431290 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.432275 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.435213 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjv6g\" (UniqueName: \"kubernetes.io/projected/c324ea9b-913d-4766-ac5a-790941b21763-kube-api-access-kjv6g\") pod \"test-operator-controller-manager-7866795846-p5tgj\" (UID: \"c324ea9b-913d-4766-ac5a-790941b21763\") " pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.435713 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-zc4ps" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.446706 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542270 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542609 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmmxq\" (UniqueName: \"kubernetes.io/projected/a67d76ca-f5c0-4416-a85b-9de9211aed83-kube-api-access-zmmxq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lk9pm\" (UID: \"a67d76ca-f5c0-4416-a85b-9de9211aed83\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542660 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cq8d\" (UniqueName: \"kubernetes.io/projected/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-kube-api-access-7cq8d\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542693 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjv6g\" (UniqueName: \"kubernetes.io/projected/c324ea9b-913d-4766-ac5a-790941b21763-kube-api-access-kjv6g\") pod \"test-operator-controller-manager-7866795846-p5tgj\" (UID: \"c324ea9b-913d-4766-ac5a-790941b21763\") " pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542718 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542745 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkmgz\" (UniqueName: \"kubernetes.io/projected/7b3527ea-31e6-4b64-b5b1-1e683fe05689-kube-api-access-hkmgz\") pod \"watcher-operator-controller-manager-5db88f68c-4h4j7\" (UID: \"7b3527ea-31e6-4b64-b5b1-1e683fe05689\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542767 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.542801 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.543058 4492 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.543114 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert podName:41543a2e-55d9-465c-bb77-f2d7b04b7e2b nodeName:}" failed. No retries permitted until 2026-02-20 06:53:03.54309457 +0000 UTC m=+740.314383547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" (UID: "41543a2e-55d9-465c-bb77-f2d7b04b7e2b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.543278 4492 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.543341 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert podName:bab88145-f203-4b8c-b9d3-a03ee1af5ac7 nodeName:}" failed. No retries permitted until 2026-02-20 06:53:03.543324262 +0000 UTC m=+740.314613240 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert") pod "infra-operator-controller-manager-79d975b745-szmrb" (UID: "bab88145-f203-4b8c-b9d3-a03ee1af5ac7") : secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.563020 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjv6g\" (UniqueName: \"kubernetes.io/projected/c324ea9b-913d-4766-ac5a-790941b21763-kube-api-access-kjv6g\") pod \"test-operator-controller-manager-7866795846-p5tgj\" (UID: \"c324ea9b-913d-4766-ac5a-790941b21763\") " pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.572529 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.644348 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmmxq\" (UniqueName: \"kubernetes.io/projected/a67d76ca-f5c0-4416-a85b-9de9211aed83-kube-api-access-zmmxq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lk9pm\" (UID: \"a67d76ca-f5c0-4416-a85b-9de9211aed83\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.644448 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cq8d\" (UniqueName: \"kubernetes.io/projected/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-kube-api-access-7cq8d\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.644516 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkmgz\" (UniqueName: \"kubernetes.io/projected/7b3527ea-31e6-4b64-b5b1-1e683fe05689-kube-api-access-hkmgz\") pod \"watcher-operator-controller-manager-5db88f68c-4h4j7\" (UID: \"7b3527ea-31e6-4b64-b5b1-1e683fe05689\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.644571 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.644613 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.646067 4492 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.646104 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:03.146091728 +0000 UTC m=+739.917380706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "metrics-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.646237 4492 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: E0220 06:53:02.646259 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:03.146253201 +0000 UTC m=+739.917542180 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "webhook-server-cert" not found Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.649220 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.659739 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmmxq\" (UniqueName: \"kubernetes.io/projected/a67d76ca-f5c0-4416-a85b-9de9211aed83-kube-api-access-zmmxq\") pod \"rabbitmq-cluster-operator-manager-668c99d594-lk9pm\" (UID: \"a67d76ca-f5c0-4416-a85b-9de9211aed83\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.666557 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cq8d\" (UniqueName: \"kubernetes.io/projected/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-kube-api-access-7cq8d\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.670585 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkmgz\" (UniqueName: \"kubernetes.io/projected/7b3527ea-31e6-4b64-b5b1-1e683fe05689-kube-api-access-hkmgz\") pod \"watcher-operator-controller-manager-5db88f68c-4h4j7\" (UID: \"7b3527ea-31e6-4b64-b5b1-1e683fe05689\") " pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.687824 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.768187 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.777195 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" Feb 20 06:53:02 crc kubenswrapper[4492]: W0220 06:53:02.828678 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c26743a_8dbb_4bdc_a2a8_430efd8070c6.slice/crio-bd0ac9aece0cad66c16b54bc629e522dfbce30810d8a0e68ac43b21358d090e3 WatchSource:0}: Error finding container bd0ac9aece0cad66c16b54bc629e522dfbce30810d8a0e68ac43b21358d090e3: Status 404 returned error can't find the container with id bd0ac9aece0cad66c16b54bc629e522dfbce30810d8a0e68ac43b21358d090e3 Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.832131 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx"] Feb 20 06:53:02 crc kubenswrapper[4492]: I0220 06:53:02.993032 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.018196 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" event={"ID":"1f1dd67b-1f84-49a3-8ffa-bba932485c90","Type":"ContainerStarted","Data":"f4bd2f05024a823468675ddabc7b63ea6adb50be6202fa186375627ccc82a0dc"} Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.028103 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" event={"ID":"16dae257-55aa-40f7-9308-182cd7ff886b","Type":"ContainerStarted","Data":"9204547d2466db955dbd894be2c70a1e46e51f7f9edfe3bfe066c6fd3aac8177"} Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.032224 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" event={"ID":"9c26743a-8dbb-4bdc-a2a8-430efd8070c6","Type":"ContainerStarted","Data":"bd0ac9aece0cad66c16b54bc629e522dfbce30810d8a0e68ac43b21358d090e3"} Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.041734 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.056505 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.070489 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.174376 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.174435 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.174703 4492 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.174757 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:04.17474045 +0000 UTC m=+740.946029429 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "webhook-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.175088 4492 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.175114 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:04.175107262 +0000 UTC m=+740.946396241 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "metrics-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.180055 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.191152 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.212095 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.250454 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.266527 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.271776 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc"] Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.288347 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c6dlc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-69f8888797-f6zqc_openstack-operators(cafb9e27-7f46-4d0d-a27f-7f60d11d5b79): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.291353 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" podUID="cafb9e27-7f46-4d0d-a27f-7f60d11d5b79" Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.316714 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76"] Feb 20 06:53:03 crc kubenswrapper[4492]: W0220 06:53:03.317871 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91e88c5d_7cb2_4ebd_9a96_d3620ea38c97.slice/crio-e29a1b49edae08387f8934cb7d328cda972d062ec613eb88cd1e93663b5a47e4 WatchSource:0}: Error finding container e29a1b49edae08387f8934cb7d328cda972d062ec613eb88cd1e93663b5a47e4: Status 404 returned error can't find the container with id e29a1b49edae08387f8934cb7d328cda972d062ec613eb88cd1e93663b5a47e4 Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.504079 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs"] Feb 20 06:53:03 crc kubenswrapper[4492]: W0220 06:53:03.510118 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc324ea9b_913d_4766_ac5a_790941b21763.slice/crio-a0b7f548686cdb3a8888849d1db9868019d92749b9f71a8af837b0eb8fef5196 WatchSource:0}: Error finding container a0b7f548686cdb3a8888849d1db9868019d92749b9f71a8af837b0eb8fef5196: Status 404 returned error can't find the container with id a0b7f548686cdb3a8888849d1db9868019d92749b9f71a8af837b0eb8fef5196 Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.512845 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kjv6g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-7866795846-p5tgj_openstack-operators(c324ea9b-913d-4766-ac5a-790941b21763): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.514016 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" podUID="c324ea9b-913d-4766-ac5a-790941b21763" Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.515875 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-7866795846-p5tgj"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.584150 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.584468 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.584497 4492 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.584557 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert podName:bab88145-f203-4b8c-b9d3-a03ee1af5ac7 nodeName:}" failed. No retries permitted until 2026-02-20 06:53:05.584539897 +0000 UTC m=+742.355828874 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert") pod "infra-operator-controller-manager-79d975b745-szmrb" (UID: "bab88145-f203-4b8c-b9d3-a03ee1af5ac7") : secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.584643 4492 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.584689 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert podName:41543a2e-55d9-465c-bb77-f2d7b04b7e2b nodeName:}" failed. No retries permitted until 2026-02-20 06:53:05.584671415 +0000 UTC m=+742.355960393 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" (UID: "41543a2e-55d9-465c-bb77-f2d7b04b7e2b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.689160 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt"] Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.704922 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jjx7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-77tpg_openstack-operators(ba9af5af-1c2b-421c-a59a-4f4a203a8e3e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.705270 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7"] Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.706052 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" podUID="ba9af5af-1c2b-421c-a59a-4f4a203a8e3e" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.709409 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fhdtl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-6hbwt_openstack-operators(10788676-0981-46c0-8a6a-eda4f457e117): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.711439 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" podUID="10788676-0981-46c0-8a6a-eda4f457e117" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.711601 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zmmxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-lk9pm_openstack-operators(a67d76ca-f5c0-4416-a85b-9de9211aed83): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 20 06:53:03 crc kubenswrapper[4492]: E0220 06:53:03.713888 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" podUID="a67d76ca-f5c0-4416-a85b-9de9211aed83" Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.715806 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-77tpg"] Feb 20 06:53:03 crc kubenswrapper[4492]: I0220 06:53:03.724962 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm"] Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.047029 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" event={"ID":"b1cfd35f-eae6-453c-8687-6bf96945edf2","Type":"ContainerStarted","Data":"00663a2052fc9f22ac571af26ef18db66fbf1d364e0294612ce48c3bdb0a88d2"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.049602 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" event={"ID":"ea530a5e-2861-4559-b80a-ef8589d9788f","Type":"ContainerStarted","Data":"d3821960aaac7fe6ffd9ff9620c391a06eaf46b868ff4e0de05391bf919fef7d"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.051136 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" event={"ID":"91e88c5d-7cb2-4ebd-9a96-d3620ea38c97","Type":"ContainerStarted","Data":"e29a1b49edae08387f8934cb7d328cda972d062ec613eb88cd1e93663b5a47e4"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.052709 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" event={"ID":"7e6bf429-fb74-4a04-ac01-b4f60d4c00c2","Type":"ContainerStarted","Data":"ff7e96e6b1a9098f166e7d02ce36296941fadd23aa00010ff18fc1ab4008a143"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.054228 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" event={"ID":"e2b03896-ea8a-469a-8a43-4f04f2108071","Type":"ContainerStarted","Data":"42e0db689607465eb3a43ea7def20d34d3049f0574b658075efd99931c4c7a4a"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.058106 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" event={"ID":"80579e11-ccf9-4377-a013-345a55ac33ab","Type":"ContainerStarted","Data":"08d4cef22e92ae82559284dbaa62665a5cdbfe264665f391a435018b190fb93b"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.059934 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" event={"ID":"d1f3b23e-c05b-42f0-af00-5af28a768c0e","Type":"ContainerStarted","Data":"ab9a20f22dbfdca8be0c254f5fed50101272d2f43f4486636d25fbe7b1f2a20a"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.061370 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" event={"ID":"cafb9e27-7f46-4d0d-a27f-7f60d11d5b79","Type":"ContainerStarted","Data":"f1703d553652b98e906c4d1e1a5dc98bd7a1ef7daca6bb4760d8b39f4439bec4"} Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.063638 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" podUID="cafb9e27-7f46-4d0d-a27f-7f60d11d5b79" Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.064562 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" event={"ID":"7b3527ea-31e6-4b64-b5b1-1e683fe05689","Type":"ContainerStarted","Data":"5a69891e5e822abfdea878aec04b8c6340cb90848fcd21bf0bc43cedee42ce21"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.072095 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" event={"ID":"86409721-594f-4c20-9f11-2f89572c3036","Type":"ContainerStarted","Data":"6979171c3a5b2be7632098331e1580ee5953f391c541db5d78c6acdc410348cf"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.078866 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" event={"ID":"a67d76ca-f5c0-4416-a85b-9de9211aed83","Type":"ContainerStarted","Data":"114fac43cf9ba8462e32671d6a6e31b8cf4dd98abe1c8ce97d172bdcb3790d25"} Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.081770 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" podUID="a67d76ca-f5c0-4416-a85b-9de9211aed83" Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.082414 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" event={"ID":"ba9af5af-1c2b-421c-a59a-4f4a203a8e3e","Type":"ContainerStarted","Data":"bbb4070a4256e26abeb961fbed944a3790ca8f39b37ba7a001b6e47100ede832"} Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.085691 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" podUID="ba9af5af-1c2b-421c-a59a-4f4a203a8e3e" Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.087234 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" event={"ID":"2918bb03-8106-4586-90bc-2618b3e78239","Type":"ContainerStarted","Data":"c49d9ed16f3c812d6f3186b93c4b6b01230964a557b2500342ed3b9a191d0247"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.095723 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" event={"ID":"883be441-1faf-4b78-9c22-16c19986902a","Type":"ContainerStarted","Data":"eef82ed3466c6901e78f35d1e876ef5759ad086c2e18aeb1e5120a0975de7fcc"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.099556 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" event={"ID":"10788676-0981-46c0-8a6a-eda4f457e117","Type":"ContainerStarted","Data":"7e9214e495258ae6cfa6000cf5382ad721d2b8ab17f0511d6fb91524c5c71b47"} Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.103951 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" podUID="10788676-0981-46c0-8a6a-eda4f457e117" Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.109265 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" event={"ID":"dd787991-93c0-4baf-8a53-c2ec22c09ede","Type":"ContainerStarted","Data":"a992d3bb20fc8ecb49448a44d65c8754f22405daf79315d9934217a9978e4aa6"} Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.122073 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" event={"ID":"c324ea9b-913d-4766-ac5a-790941b21763","Type":"ContainerStarted","Data":"a0b7f548686cdb3a8888849d1db9868019d92749b9f71a8af837b0eb8fef5196"} Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.125424 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" podUID="c324ea9b-913d-4766-ac5a-790941b21763" Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.199794 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.199987 4492 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.200061 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:06.200039113 +0000 UTC m=+742.971328092 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "metrics-server-cert" not found Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.200069 4492 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 06:53:04 crc kubenswrapper[4492]: E0220 06:53:04.200143 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:06.200128432 +0000 UTC m=+742.971417400 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "webhook-server-cert" not found Feb 20 06:53:04 crc kubenswrapper[4492]: I0220 06:53:04.199913 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.151519 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f0fabdf79095def0f8b1c0442925548a94ca94bed4de2d3b171277129f8079e6\\\"\"" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" podUID="c324ea9b-913d-4766-ac5a-790941b21763" Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.151505 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" podUID="a67d76ca-f5c0-4416-a85b-9de9211aed83" Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.151551 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:229fc8c8d94dd4102d2151cd4ec1eaaa09d897c2b396d06e903f61ea29c1fa34\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" podUID="cafb9e27-7f46-4d0d-a27f-7f60d11d5b79" Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.151933 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" podUID="ba9af5af-1c2b-421c-a59a-4f4a203a8e3e" Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.156944 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" podUID="10788676-0981-46c0-8a6a-eda4f457e117" Feb 20 06:53:05 crc kubenswrapper[4492]: I0220 06:53:05.626802 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:05 crc kubenswrapper[4492]: I0220 06:53:05.626861 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.627053 4492 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.627101 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert podName:41543a2e-55d9-465c-bb77-f2d7b04b7e2b nodeName:}" failed. No retries permitted until 2026-02-20 06:53:09.627086035 +0000 UTC m=+746.398375014 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" (UID: "41543a2e-55d9-465c-bb77-f2d7b04b7e2b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.627309 4492 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:05 crc kubenswrapper[4492]: E0220 06:53:05.627437 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert podName:bab88145-f203-4b8c-b9d3-a03ee1af5ac7 nodeName:}" failed. No retries permitted until 2026-02-20 06:53:09.627405617 +0000 UTC m=+746.398694596 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert") pod "infra-operator-controller-manager-79d975b745-szmrb" (UID: "bab88145-f203-4b8c-b9d3-a03ee1af5ac7") : secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:06 crc kubenswrapper[4492]: I0220 06:53:06.242132 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:06 crc kubenswrapper[4492]: I0220 06:53:06.242188 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:06 crc kubenswrapper[4492]: E0220 06:53:06.242345 4492 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 06:53:06 crc kubenswrapper[4492]: E0220 06:53:06.242390 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:10.242376059 +0000 UTC m=+747.013665036 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "webhook-server-cert" not found Feb 20 06:53:06 crc kubenswrapper[4492]: E0220 06:53:06.242913 4492 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 06:53:06 crc kubenswrapper[4492]: E0220 06:53:06.243006 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:10.242985626 +0000 UTC m=+747.014274605 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "metrics-server-cert" not found Feb 20 06:53:09 crc kubenswrapper[4492]: I0220 06:53:09.311325 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:53:09 crc kubenswrapper[4492]: I0220 06:53:09.311409 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:53:09 crc kubenswrapper[4492]: I0220 06:53:09.311489 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:53:09 crc kubenswrapper[4492]: I0220 06:53:09.312190 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5090e1416f91468b7a37c5f3132c069b8613e7ce0e4c8c57bc77b7ef73cc9bce"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 06:53:09 crc kubenswrapper[4492]: I0220 06:53:09.312252 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://5090e1416f91468b7a37c5f3132c069b8613e7ce0e4c8c57bc77b7ef73cc9bce" gracePeriod=600 Feb 20 06:53:09 crc kubenswrapper[4492]: I0220 06:53:09.697794 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:09 crc kubenswrapper[4492]: I0220 06:53:09.698257 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:09 crc kubenswrapper[4492]: E0220 06:53:09.697896 4492 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:09 crc kubenswrapper[4492]: E0220 06:53:09.698499 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert podName:bab88145-f203-4b8c-b9d3-a03ee1af5ac7 nodeName:}" failed. No retries permitted until 2026-02-20 06:53:17.698467991 +0000 UTC m=+754.469756969 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert") pod "infra-operator-controller-manager-79d975b745-szmrb" (UID: "bab88145-f203-4b8c-b9d3-a03ee1af5ac7") : secret "infra-operator-webhook-server-cert" not found Feb 20 06:53:09 crc kubenswrapper[4492]: E0220 06:53:09.698345 4492 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:09 crc kubenswrapper[4492]: E0220 06:53:09.698545 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert podName:41543a2e-55d9-465c-bb77-f2d7b04b7e2b nodeName:}" failed. No retries permitted until 2026-02-20 06:53:17.698528184 +0000 UTC m=+754.469817163 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" (UID: "41543a2e-55d9-465c-bb77-f2d7b04b7e2b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:10 crc kubenswrapper[4492]: I0220 06:53:10.189126 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="5090e1416f91468b7a37c5f3132c069b8613e7ce0e4c8c57bc77b7ef73cc9bce" exitCode=0 Feb 20 06:53:10 crc kubenswrapper[4492]: I0220 06:53:10.189183 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"5090e1416f91468b7a37c5f3132c069b8613e7ce0e4c8c57bc77b7ef73cc9bce"} Feb 20 06:53:10 crc kubenswrapper[4492]: I0220 06:53:10.189259 4492 scope.go:117] "RemoveContainer" containerID="855da3473a6d9146a28853331dfc686fe2c8735ee6afb03897c4083de78fb717" Feb 20 06:53:10 crc kubenswrapper[4492]: I0220 06:53:10.307650 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:10 crc kubenswrapper[4492]: I0220 06:53:10.307733 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:10 crc kubenswrapper[4492]: E0220 06:53:10.307990 4492 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 06:53:10 crc kubenswrapper[4492]: E0220 06:53:10.308065 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:18.308044464 +0000 UTC m=+755.079333442 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "webhook-server-cert" not found Feb 20 06:53:10 crc kubenswrapper[4492]: E0220 06:53:10.308527 4492 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 20 06:53:10 crc kubenswrapper[4492]: E0220 06:53:10.308578 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:18.308568401 +0000 UTC m=+755.079857380 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "metrics-server-cert" not found Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.150086 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:e8a675284ff97a1d3f0f07583863be20b20b4aa48ebb34dbc80d83fe39d757b2" Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.151264 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:e8a675284ff97a1d3f0f07583863be20b20b4aa48ebb34dbc80d83fe39d757b2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6pkq2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-69f49c598c-5kqdb_openstack-operators(80579e11-ccf9-4377-a013-345a55ac33ab): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.152498 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" podUID="80579e11-ccf9-4377-a013-345a55ac33ab" Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.264537 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:e8a675284ff97a1d3f0f07583863be20b20b4aa48ebb34dbc80d83fe39d757b2\\\"\"" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" podUID="80579e11-ccf9-4377-a013-345a55ac33ab" Feb 20 06:53:17 crc kubenswrapper[4492]: I0220 06:53:17.729269 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:17 crc kubenswrapper[4492]: I0220 06:53:17.729342 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.729565 4492 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.729641 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert podName:41543a2e-55d9-465c-bb77-f2d7b04b7e2b nodeName:}" failed. No retries permitted until 2026-02-20 06:53:33.729619416 +0000 UTC m=+770.500908393 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert") pod "openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" (UID: "41543a2e-55d9-465c-bb77-f2d7b04b7e2b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 20 06:53:17 crc kubenswrapper[4492]: I0220 06:53:17.737212 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bab88145-f203-4b8c-b9d3-a03ee1af5ac7-cert\") pod \"infra-operator-controller-manager-79d975b745-szmrb\" (UID: \"bab88145-f203-4b8c-b9d3-a03ee1af5ac7\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:17 crc kubenswrapper[4492]: I0220 06:53:17.929386 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.947459 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838" Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.947723 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b5g76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-2zzhs_openstack-operators(883be441-1faf-4b78-9c22-16c19986902a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:53:17 crc kubenswrapper[4492]: E0220 06:53:17.949647 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" podUID="883be441-1faf-4b78-9c22-16c19986902a" Feb 20 06:53:18 crc kubenswrapper[4492]: E0220 06:53:18.271564 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" podUID="883be441-1faf-4b78-9c22-16c19986902a" Feb 20 06:53:18 crc kubenswrapper[4492]: I0220 06:53:18.339129 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:18 crc kubenswrapper[4492]: I0220 06:53:18.339200 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:18 crc kubenswrapper[4492]: E0220 06:53:18.339493 4492 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 20 06:53:18 crc kubenswrapper[4492]: E0220 06:53:18.339617 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs podName:3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf nodeName:}" failed. No retries permitted until 2026-02-20 06:53:34.339588939 +0000 UTC m=+771.110877916 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs") pod "openstack-operator-controller-manager-69ff7bc449-242g8" (UID: "3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf") : secret "webhook-server-cert" not found Feb 20 06:53:18 crc kubenswrapper[4492]: I0220 06:53:18.344921 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-metrics-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:18 crc kubenswrapper[4492]: E0220 06:53:18.535758 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867" Feb 20 06:53:18 crc kubenswrapper[4492]: E0220 06:53:18.536013 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-58r6t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-554564d7fc-s7bxs_openstack-operators(86409721-594f-4c20-9f11-2f89572c3036): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:53:18 crc kubenswrapper[4492]: E0220 06:53:18.537556 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" podUID="86409721-594f-4c20-9f11-2f89572c3036" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.210278 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.211091 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2lgf8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-6d8bf5c495-h27fh_openstack-operators(16dae257-55aa-40f7-9308-182cd7ff886b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.212884 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" podUID="16dae257-55aa-40f7-9308-182cd7ff886b" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.277354 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642\\\"\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" podUID="16dae257-55aa-40f7-9308-182cd7ff886b" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.277752 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" podUID="86409721-594f-4c20-9f11-2f89572c3036" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.887916 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.888204 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4c2g7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-5d946d989d-vqxgx_openstack-operators(1f1dd67b-1f84-49a3-8ffa-bba932485c90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:53:19 crc kubenswrapper[4492]: E0220 06:53:19.889402 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" podUID="1f1dd67b-1f84-49a3-8ffa-bba932485c90" Feb 20 06:53:20 crc kubenswrapper[4492]: E0220 06:53:20.296972 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:2b8ab3063af4aaeed0198197aae6f391c6647ac686c94c85668537f1d5933979\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" podUID="1f1dd67b-1f84-49a3-8ffa-bba932485c90" Feb 20 06:53:20 crc kubenswrapper[4492]: E0220 06:53:20.767804 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99" Feb 20 06:53:20 crc kubenswrapper[4492]: E0220 06:53:20.767987 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s8pb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-7f45b4ff68-t4h7v_openstack-operators(e2b03896-ea8a-469a-8a43-4f04f2108071): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:53:20 crc kubenswrapper[4492]: E0220 06:53:20.769119 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" podUID="e2b03896-ea8a-469a-8a43-4f04f2108071" Feb 20 06:53:21 crc kubenswrapper[4492]: E0220 06:53:21.316026 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:66a4b9322ebb573313178ea88e31026d4532f461592b9fae2dff71efd9256d99\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" podUID="e2b03896-ea8a-469a-8a43-4f04f2108071" Feb 20 06:53:21 crc kubenswrapper[4492]: E0220 06:53:21.451603 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1" Feb 20 06:53:21 crc kubenswrapper[4492]: E0220 06:53:21.451841 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bbkkl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b4d948c87-t8cd6_openstack-operators(2918bb03-8106-4586-90bc-2618b3e78239): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:53:21 crc kubenswrapper[4492]: E0220 06:53:21.453036 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" podUID="2918bb03-8106-4586-90bc-2618b3e78239" Feb 20 06:53:22 crc kubenswrapper[4492]: E0220 06:53:22.321241 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" podUID="2918bb03-8106-4586-90bc-2618b3e78239" Feb 20 06:53:23 crc kubenswrapper[4492]: I0220 06:53:23.330161 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"721ada477bd6e659aca9b99823e1dcf15d8562664d9f214bcd56e2ac3871cc6b"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.081379 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-szmrb"] Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.365637 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" event={"ID":"10788676-0981-46c0-8a6a-eda4f457e117","Type":"ContainerStarted","Data":"de2c952aadd4363ceccba6d02f0ba6760213c68861c9ff918c152898da3c3150"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.366254 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.367448 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" event={"ID":"9c26743a-8dbb-4bdc-a2a8-430efd8070c6","Type":"ContainerStarted","Data":"c7fb3ebabb3f87860868c5aec1952b7471f77e11f653ed058b86247e75d9128a"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.367813 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.372460 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" event={"ID":"d1f3b23e-c05b-42f0-af00-5af28a768c0e","Type":"ContainerStarted","Data":"2271ea9860edaebe92af45fd8705d9d35bdce36f5e518cb3f12bdddf8dfae4a5"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.372840 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.377240 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" event={"ID":"b1cfd35f-eae6-453c-8687-6bf96945edf2","Type":"ContainerStarted","Data":"cc8eda9d90b3445490767c4fa19dd1e3f5716a0214f7fbae780c79f898dc81f0"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.377581 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.378706 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" event={"ID":"7b3527ea-31e6-4b64-b5b1-1e683fe05689","Type":"ContainerStarted","Data":"daf977b67c6e3e54ac4338d187685bfa5b7b62ac439cacbed41659b25d4497b1"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.379018 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.379939 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" event={"ID":"91e88c5d-7cb2-4ebd-9a96-d3620ea38c97","Type":"ContainerStarted","Data":"f9b04d3057ba64bce68f9d8439b13e3f969bb78b7b16677b0f677b4cf57eebd9"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.380245 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.381216 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" event={"ID":"7e6bf429-fb74-4a04-ac01-b4f60d4c00c2","Type":"ContainerStarted","Data":"00e1871b244d96f13faa99a737273c9df8da85eaa054866a8bbdd5b90d4bb95c"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.381559 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.392158 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" podStartSLOduration=4.487773767 podStartE2EDuration="27.392146615s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.709306056 +0000 UTC m=+740.480595034" lastFinishedPulling="2026-02-20 06:53:26.613678905 +0000 UTC m=+763.384967882" observedRunningTime="2026-02-20 06:53:28.388253899 +0000 UTC m=+765.159542876" watchObservedRunningTime="2026-02-20 06:53:28.392146615 +0000 UTC m=+765.163435594" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.432344 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" event={"ID":"dd787991-93c0-4baf-8a53-c2ec22c09ede","Type":"ContainerStarted","Data":"9e9fb8116db5bb1d1d3f1edf5c8fef5e32c247f1b2f47c71b8ea28c3acf19309"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.435886 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.448418 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" event={"ID":"c324ea9b-913d-4766-ac5a-790941b21763","Type":"ContainerStarted","Data":"23517a7696352047a4a78f8a1e7b3cc255ba9e180b1cdc764b3f65bcdc8769ae"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.449043 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.484735 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" event={"ID":"ea530a5e-2861-4559-b80a-ef8589d9788f","Type":"ContainerStarted","Data":"5ff1b86a3610a70534a726714a76f3db380da126c2cf0f13afc9e4f3d2838d0d"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.485305 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.497949 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" podStartSLOduration=8.609994477 podStartE2EDuration="27.497926488s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.068085057 +0000 UTC m=+739.839374035" lastFinishedPulling="2026-02-20 06:53:21.956017068 +0000 UTC m=+758.727306046" observedRunningTime="2026-02-20 06:53:28.428433803 +0000 UTC m=+765.199722781" watchObservedRunningTime="2026-02-20 06:53:28.497926488 +0000 UTC m=+765.269215467" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.520219 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" event={"ID":"bab88145-f203-4b8c-b9d3-a03ee1af5ac7","Type":"ContainerStarted","Data":"4860f3236b921c666c72ecfbf61ab95dc6b7599c0c31247df6c5bb31db02b921"} Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.564128 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" podStartSLOduration=8.310108886 podStartE2EDuration="26.564103776s" podCreationTimestamp="2026-02-20 06:53:02 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.702069656 +0000 UTC m=+740.473358635" lastFinishedPulling="2026-02-20 06:53:21.956064548 +0000 UTC m=+758.727353525" observedRunningTime="2026-02-20 06:53:28.46172349 +0000 UTC m=+765.233012468" watchObservedRunningTime="2026-02-20 06:53:28.564103776 +0000 UTC m=+765.335392754" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.572907 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" podStartSLOduration=8.905086413 podStartE2EDuration="27.572890026s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.286859904 +0000 UTC m=+740.058148882" lastFinishedPulling="2026-02-20 06:53:21.954663517 +0000 UTC m=+758.725952495" observedRunningTime="2026-02-20 06:53:28.516456526 +0000 UTC m=+765.287745524" watchObservedRunningTime="2026-02-20 06:53:28.572890026 +0000 UTC m=+765.344179004" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.579626 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" podStartSLOduration=8.521598834 podStartE2EDuration="27.579609803s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:02.897286291 +0000 UTC m=+739.668575260" lastFinishedPulling="2026-02-20 06:53:21.955297251 +0000 UTC m=+758.726586229" observedRunningTime="2026-02-20 06:53:28.556917039 +0000 UTC m=+765.328206017" watchObservedRunningTime="2026-02-20 06:53:28.579609803 +0000 UTC m=+765.350898781" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.595835 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" podStartSLOduration=8.887223935 podStartE2EDuration="27.595818214s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.246052025 +0000 UTC m=+740.017341003" lastFinishedPulling="2026-02-20 06:53:21.954646303 +0000 UTC m=+758.725935282" observedRunningTime="2026-02-20 06:53:28.587068602 +0000 UTC m=+765.358357570" watchObservedRunningTime="2026-02-20 06:53:28.595818214 +0000 UTC m=+765.367107192" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.611439 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" podStartSLOduration=7.95953219 podStartE2EDuration="27.611426784s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.320665082 +0000 UTC m=+740.091954060" lastFinishedPulling="2026-02-20 06:53:22.972559675 +0000 UTC m=+759.743848654" observedRunningTime="2026-02-20 06:53:28.608282869 +0000 UTC m=+765.379571846" watchObservedRunningTime="2026-02-20 06:53:28.611426784 +0000 UTC m=+765.382715763" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.669164 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" podStartSLOduration=8.834937101 podStartE2EDuration="27.669145608s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.121658518 +0000 UTC m=+739.892947486" lastFinishedPulling="2026-02-20 06:53:21.955867025 +0000 UTC m=+758.727155993" observedRunningTime="2026-02-20 06:53:28.666004086 +0000 UTC m=+765.437293064" watchObservedRunningTime="2026-02-20 06:53:28.669145608 +0000 UTC m=+765.440434586" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.669806 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" podStartSLOduration=3.250208176 podStartE2EDuration="27.669797737s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.512672735 +0000 UTC m=+740.283961713" lastFinishedPulling="2026-02-20 06:53:27.932262306 +0000 UTC m=+764.703551274" observedRunningTime="2026-02-20 06:53:28.650490606 +0000 UTC m=+765.421779594" watchObservedRunningTime="2026-02-20 06:53:28.669797737 +0000 UTC m=+765.441086715" Feb 20 06:53:28 crc kubenswrapper[4492]: I0220 06:53:28.696812 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" podStartSLOduration=8.784938139 podStartE2EDuration="27.696783212s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.044060494 +0000 UTC m=+739.815349472" lastFinishedPulling="2026-02-20 06:53:21.955905568 +0000 UTC m=+758.727194545" observedRunningTime="2026-02-20 06:53:28.695510593 +0000 UTC m=+765.466799571" watchObservedRunningTime="2026-02-20 06:53:28.696783212 +0000 UTC m=+765.468072180" Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.528670 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" event={"ID":"ba9af5af-1c2b-421c-a59a-4f4a203a8e3e","Type":"ContainerStarted","Data":"69f4a7e975411be1641e86f4b36ee35bbfdcd82976bf84450c37d76b0bfc114b"} Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.529251 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.530614 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" event={"ID":"cafb9e27-7f46-4d0d-a27f-7f60d11d5b79","Type":"ContainerStarted","Data":"06f6242b74c3127577bf1179c9111c6833d289bc52f80a0f4cbf3e125556603c"} Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.530880 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.532218 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" event={"ID":"a67d76ca-f5c0-4416-a85b-9de9211aed83","Type":"ContainerStarted","Data":"ed0ee8a5845f93cd3afb68b7f9ecfe071091b357cfd8c3d168798ea6c8d8d7cd"} Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.572054 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" podStartSLOduration=4.342564104 podStartE2EDuration="28.572035604s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.704789263 +0000 UTC m=+740.476078241" lastFinishedPulling="2026-02-20 06:53:27.934260763 +0000 UTC m=+764.705549741" observedRunningTime="2026-02-20 06:53:29.555250196 +0000 UTC m=+766.326539173" watchObservedRunningTime="2026-02-20 06:53:29.572035604 +0000 UTC m=+766.343324582" Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.572549 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" podStartSLOduration=3.915034891 podStartE2EDuration="28.572541798s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.288207603 +0000 UTC m=+740.059496582" lastFinishedPulling="2026-02-20 06:53:27.945714511 +0000 UTC m=+764.717003489" observedRunningTime="2026-02-20 06:53:29.56796467 +0000 UTC m=+766.339253648" watchObservedRunningTime="2026-02-20 06:53:29.572541798 +0000 UTC m=+766.343830776" Feb 20 06:53:29 crc kubenswrapper[4492]: I0220 06:53:29.581176 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-lk9pm" podStartSLOduration=3.360465618 podStartE2EDuration="27.581168248s" podCreationTimestamp="2026-02-20 06:53:02 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.711399944 +0000 UTC m=+740.482688922" lastFinishedPulling="2026-02-20 06:53:27.932102574 +0000 UTC m=+764.703391552" observedRunningTime="2026-02-20 06:53:29.578546175 +0000 UTC m=+766.349835143" watchObservedRunningTime="2026-02-20 06:53:29.581168248 +0000 UTC m=+766.352457225" Feb 20 06:53:31 crc kubenswrapper[4492]: I0220 06:53:31.545996 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" event={"ID":"883be441-1faf-4b78-9c22-16c19986902a","Type":"ContainerStarted","Data":"860bcdca5155aaa43b26292eec24623efe90fa04f8e257df455b2352df7b99d7"} Feb 20 06:53:31 crc kubenswrapper[4492]: I0220 06:53:31.547215 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" Feb 20 06:53:31 crc kubenswrapper[4492]: I0220 06:53:31.548734 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" event={"ID":"bab88145-f203-4b8c-b9d3-a03ee1af5ac7","Type":"ContainerStarted","Data":"654fca298d0907132d52ef8275484aaa76af5881ed82681dba20023b781a5c56"} Feb 20 06:53:31 crc kubenswrapper[4492]: I0220 06:53:31.548958 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:31 crc kubenswrapper[4492]: I0220 06:53:31.565283 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" podStartSLOduration=2.477016317 podStartE2EDuration="30.565266639s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.229832133 +0000 UTC m=+740.001121111" lastFinishedPulling="2026-02-20 06:53:31.318082465 +0000 UTC m=+768.089371433" observedRunningTime="2026-02-20 06:53:31.563380093 +0000 UTC m=+768.334669061" watchObservedRunningTime="2026-02-20 06:53:31.565266639 +0000 UTC m=+768.336555617" Feb 20 06:53:31 crc kubenswrapper[4492]: I0220 06:53:31.601852 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" podStartSLOduration=27.390102724 podStartE2EDuration="30.601833574s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:28.109099784 +0000 UTC m=+764.880388762" lastFinishedPulling="2026-02-20 06:53:31.320830635 +0000 UTC m=+768.092119612" observedRunningTime="2026-02-20 06:53:31.598773596 +0000 UTC m=+768.370062574" watchObservedRunningTime="2026-02-20 06:53:31.601833574 +0000 UTC m=+768.373122541" Feb 20 06:53:32 crc kubenswrapper[4492]: I0220 06:53:32.556837 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" event={"ID":"86409721-594f-4c20-9f11-2f89572c3036","Type":"ContainerStarted","Data":"ef6fc6c69f82a180ec9e32d561d0e2a36bb158e94e8ea1b66331595d78adc03e"} Feb 20 06:53:32 crc kubenswrapper[4492]: I0220 06:53:32.557449 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" Feb 20 06:53:32 crc kubenswrapper[4492]: I0220 06:53:32.562279 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" event={"ID":"80579e11-ccf9-4377-a013-345a55ac33ab","Type":"ContainerStarted","Data":"6c4afcb6c2d012be39debcc9552c61c91c13697c003d30866dbebed12a5b4b6f"} Feb 20 06:53:32 crc kubenswrapper[4492]: I0220 06:53:32.562512 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" Feb 20 06:53:32 crc kubenswrapper[4492]: I0220 06:53:32.585756 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" podStartSLOduration=3.073936856 podStartE2EDuration="31.585737221s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.504843668 +0000 UTC m=+740.276132647" lastFinishedPulling="2026-02-20 06:53:32.016644034 +0000 UTC m=+768.787933012" observedRunningTime="2026-02-20 06:53:32.583532245 +0000 UTC m=+769.354821223" watchObservedRunningTime="2026-02-20 06:53:32.585737221 +0000 UTC m=+769.357026199" Feb 20 06:53:32 crc kubenswrapper[4492]: I0220 06:53:32.599681 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" podStartSLOduration=2.574974007 podStartE2EDuration="31.599665795s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.043245357 +0000 UTC m=+739.814534336" lastFinishedPulling="2026-02-20 06:53:32.067937146 +0000 UTC m=+768.839226124" observedRunningTime="2026-02-20 06:53:32.594516168 +0000 UTC m=+769.365805146" watchObservedRunningTime="2026-02-20 06:53:32.599665795 +0000 UTC m=+769.370954773" Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.580410 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" event={"ID":"e2b03896-ea8a-469a-8a43-4f04f2108071","Type":"ContainerStarted","Data":"6fdbbdb6bdb9d38425c9b9a7610e878acc94feb765567d154186b9e0b00da2a1"} Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.584261 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.586856 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" event={"ID":"16dae257-55aa-40f7-9308-182cd7ff886b","Type":"ContainerStarted","Data":"2cf0d6631fd77957ea8b0cbf8ca236bea6cee503e1fc0907a9409ae47fec88a4"} Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.614506 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" podStartSLOduration=2.881380778 podStartE2EDuration="32.614485896s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.28725093 +0000 UTC m=+740.058539908" lastFinishedPulling="2026-02-20 06:53:33.020356049 +0000 UTC m=+769.791645026" observedRunningTime="2026-02-20 06:53:33.603522862 +0000 UTC m=+770.374811840" watchObservedRunningTime="2026-02-20 06:53:33.614485896 +0000 UTC m=+770.385774874" Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.631668 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" podStartSLOduration=2.243037302 podStartE2EDuration="32.63165085s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:02.690844983 +0000 UTC m=+739.462133960" lastFinishedPulling="2026-02-20 06:53:33.07945854 +0000 UTC m=+769.850747508" observedRunningTime="2026-02-20 06:53:33.628045204 +0000 UTC m=+770.399334182" watchObservedRunningTime="2026-02-20 06:53:33.63165085 +0000 UTC m=+770.402939829" Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.737350 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.746894 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/41543a2e-55d9-465c-bb77-f2d7b04b7e2b-cert\") pod \"openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv\" (UID: \"41543a2e-55d9-465c-bb77-f2d7b04b7e2b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:33 crc kubenswrapper[4492]: I0220 06:53:33.755588 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.025948 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv"] Feb 20 06:53:34 crc kubenswrapper[4492]: W0220 06:53:34.042039 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41543a2e_55d9_465c_bb77_f2d7b04b7e2b.slice/crio-a5d276ab170ddc2cc4cfe7877cf610f6248c099384ea9995972d07ee6c95e814 WatchSource:0}: Error finding container a5d276ab170ddc2cc4cfe7877cf610f6248c099384ea9995972d07ee6c95e814: Status 404 returned error can't find the container with id a5d276ab170ddc2cc4cfe7877cf610f6248c099384ea9995972d07ee6c95e814 Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.360546 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.366370 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf-webhook-certs\") pod \"openstack-operator-controller-manager-69ff7bc449-242g8\" (UID: \"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf\") " pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.520666 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.599049 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" event={"ID":"2918bb03-8106-4586-90bc-2618b3e78239","Type":"ContainerStarted","Data":"7b21298ee654a919686fe6dd617dae2596bff26dbfe458a06f86487289c2465c"} Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.599331 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.602713 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" event={"ID":"41543a2e-55d9-465c-bb77-f2d7b04b7e2b","Type":"ContainerStarted","Data":"a5d276ab170ddc2cc4cfe7877cf610f6248c099384ea9995972d07ee6c95e814"} Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.633087 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" podStartSLOduration=2.803470987 podStartE2EDuration="33.633071425s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:03.228693396 +0000 UTC m=+739.999982374" lastFinishedPulling="2026-02-20 06:53:34.058293834 +0000 UTC m=+770.829582812" observedRunningTime="2026-02-20 06:53:34.631656698 +0000 UTC m=+771.402945676" watchObservedRunningTime="2026-02-20 06:53:34.633071425 +0000 UTC m=+771.404360403" Feb 20 06:53:34 crc kubenswrapper[4492]: I0220 06:53:34.952818 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8"] Feb 20 06:53:34 crc kubenswrapper[4492]: W0220 06:53:34.954912 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bd31619_5d0d_4da6_bc6d_3f7fab1a67cf.slice/crio-5a6bd7325891b8152f249ed6b58bcfba5035c8fff5418c09cd5f0b062cc7d721 WatchSource:0}: Error finding container 5a6bd7325891b8152f249ed6b58bcfba5035c8fff5418c09cd5f0b062cc7d721: Status 404 returned error can't find the container with id 5a6bd7325891b8152f249ed6b58bcfba5035c8fff5418c09cd5f0b062cc7d721 Feb 20 06:53:35 crc kubenswrapper[4492]: I0220 06:53:35.618717 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" event={"ID":"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf","Type":"ContainerStarted","Data":"0d4fff04ecf082e42c9c65246051f01163d813875b6203c8be08c419bf2d158b"} Feb 20 06:53:35 crc kubenswrapper[4492]: I0220 06:53:35.618797 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" event={"ID":"3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf","Type":"ContainerStarted","Data":"5a6bd7325891b8152f249ed6b58bcfba5035c8fff5418c09cd5f0b062cc7d721"} Feb 20 06:53:35 crc kubenswrapper[4492]: I0220 06:53:35.654010 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" podStartSLOduration=33.653993668 podStartE2EDuration="33.653993668s" podCreationTimestamp="2026-02-20 06:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:53:35.644247567 +0000 UTC m=+772.415536536" watchObservedRunningTime="2026-02-20 06:53:35.653993668 +0000 UTC m=+772.425282646" Feb 20 06:53:36 crc kubenswrapper[4492]: I0220 06:53:36.633134 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" event={"ID":"1f1dd67b-1f84-49a3-8ffa-bba932485c90","Type":"ContainerStarted","Data":"8d9d10581479426eec1256b97ecb56543bbe2bcf1353857ff66da50a6185c734"} Feb 20 06:53:36 crc kubenswrapper[4492]: I0220 06:53:36.633567 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:36 crc kubenswrapper[4492]: I0220 06:53:36.633784 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" Feb 20 06:53:36 crc kubenswrapper[4492]: I0220 06:53:36.658495 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" podStartSLOduration=2.578529217 podStartE2EDuration="35.658450356s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:02.988328445 +0000 UTC m=+739.759617423" lastFinishedPulling="2026-02-20 06:53:36.068249584 +0000 UTC m=+772.839538562" observedRunningTime="2026-02-20 06:53:36.655145065 +0000 UTC m=+773.426434044" watchObservedRunningTime="2026-02-20 06:53:36.658450356 +0000 UTC m=+773.429739334" Feb 20 06:53:37 crc kubenswrapper[4492]: I0220 06:53:37.937149 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-szmrb" Feb 20 06:53:38 crc kubenswrapper[4492]: I0220 06:53:38.650379 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" event={"ID":"41543a2e-55d9-465c-bb77-f2d7b04b7e2b","Type":"ContainerStarted","Data":"291c2a8f102c846b36377341b5eb7f7f7fc9dcf1b1040a9b4dcb4e29c19efc64"} Feb 20 06:53:38 crc kubenswrapper[4492]: I0220 06:53:38.650519 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:38 crc kubenswrapper[4492]: I0220 06:53:38.677011 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" podStartSLOduration=34.043657503 podStartE2EDuration="37.676992319s" podCreationTimestamp="2026-02-20 06:53:01 +0000 UTC" firstStartedPulling="2026-02-20 06:53:34.044725288 +0000 UTC m=+770.816014257" lastFinishedPulling="2026-02-20 06:53:37.678060095 +0000 UTC m=+774.449349073" observedRunningTime="2026-02-20 06:53:38.670849541 +0000 UTC m=+775.442138519" watchObservedRunningTime="2026-02-20 06:53:38.676992319 +0000 UTC m=+775.448281298" Feb 20 06:53:41 crc kubenswrapper[4492]: I0220 06:53:41.832046 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5d946d989d-vqxgx" Feb 20 06:53:41 crc kubenswrapper[4492]: I0220 06:53:41.845662 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" Feb 20 06:53:41 crc kubenswrapper[4492]: I0220 06:53:41.847403 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-h27fh" Feb 20 06:53:41 crc kubenswrapper[4492]: I0220 06:53:41.875581 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987464f4-f5lt2" Feb 20 06:53:41 crc kubenswrapper[4492]: I0220 06:53:41.969867 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-5kqdb" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.006553 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-jm5rz" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.058834 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-54f6768c69-dntvf" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.076604 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-64ddbf8bb-vckjp" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.106530 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-qrmfx" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.106602 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-dnt2c" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.108282 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-2zzhs" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.243926 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f8888797-f6zqc" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.279097 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-d44cf6b75-t5t76" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.309784 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-6hbwt" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.322852 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-s7bxs" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.352877 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-t8cd6" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.369528 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-77tpg" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.411431 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7f45b4ff68-t4h7v" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.576129 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-7866795846-p5tgj" Feb 20 06:53:42 crc kubenswrapper[4492]: I0220 06:53:42.691345 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5db88f68c-4h4j7" Feb 20 06:53:43 crc kubenswrapper[4492]: I0220 06:53:43.760987 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv" Feb 20 06:53:44 crc kubenswrapper[4492]: I0220 06:53:44.527587 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-69ff7bc449-242g8" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.804755 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c5669d55c-5n8bd"] Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.806315 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.810699 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.811558 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.811900 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ccgtm" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.814659 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.831925 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5669d55c-5n8bd"] Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.856642 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts7q2\" (UniqueName: \"kubernetes.io/projected/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-kube-api-access-ts7q2\") pod \"dnsmasq-dns-6c5669d55c-5n8bd\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.856741 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-config\") pod \"dnsmasq-dns-6c5669d55c-5n8bd\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.875968 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68c87b565c-qrkbl"] Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.882867 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.885416 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.925825 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68c87b565c-qrkbl"] Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.957611 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-config\") pod \"dnsmasq-dns-6c5669d55c-5n8bd\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.957707 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-config\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.957733 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpl9v\" (UniqueName: \"kubernetes.io/projected/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-kube-api-access-tpl9v\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.957770 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts7q2\" (UniqueName: \"kubernetes.io/projected/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-kube-api-access-ts7q2\") pod \"dnsmasq-dns-6c5669d55c-5n8bd\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.957803 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-dns-svc\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.958878 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-config\") pod \"dnsmasq-dns-6c5669d55c-5n8bd\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:53:59 crc kubenswrapper[4492]: I0220 06:53:59.976273 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts7q2\" (UniqueName: \"kubernetes.io/projected/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-kube-api-access-ts7q2\") pod \"dnsmasq-dns-6c5669d55c-5n8bd\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.059039 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-config\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.060115 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpl9v\" (UniqueName: \"kubernetes.io/projected/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-kube-api-access-tpl9v\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.060515 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-dns-svc\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.061176 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-dns-svc\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.060039 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-config\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.085975 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpl9v\" (UniqueName: \"kubernetes.io/projected/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-kube-api-access-tpl9v\") pod \"dnsmasq-dns-68c87b565c-qrkbl\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.122496 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.227984 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.551338 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5669d55c-5n8bd"] Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.623661 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68c87b565c-qrkbl"] Feb 20 06:54:00 crc kubenswrapper[4492]: W0220 06:54:00.627542 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d7bd976_285d_4222_b9c7_37c27bd6fd8a.slice/crio-22ba3663934c47860afc326655811585b5a4f829bbf9685b6c6254a94fcd05cc WatchSource:0}: Error finding container 22ba3663934c47860afc326655811585b5a4f829bbf9685b6c6254a94fcd05cc: Status 404 returned error can't find the container with id 22ba3663934c47860afc326655811585b5a4f829bbf9685b6c6254a94fcd05cc Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.821082 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" event={"ID":"c02da0cb-f83e-468b-b8e0-2bfc7126da7f","Type":"ContainerStarted","Data":"8ff24cf66b682f30cb30467f5fea7864cb43b6a018f429b156e49c3e06823972"} Feb 20 06:54:00 crc kubenswrapper[4492]: I0220 06:54:00.823848 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" event={"ID":"9d7bd976-285d-4222-b9c7-37c27bd6fd8a","Type":"ContainerStarted","Data":"22ba3663934c47860afc326655811585b5a4f829bbf9685b6c6254a94fcd05cc"} Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.626733 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68c87b565c-qrkbl"] Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.657445 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b46594d6f-72nrw"] Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.658652 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.679374 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b46594d6f-72nrw"] Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.705412 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-config\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.705498 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-dns-svc\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.705588 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8qfw\" (UniqueName: \"kubernetes.io/projected/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-kube-api-access-c8qfw\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.807455 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8qfw\" (UniqueName: \"kubernetes.io/projected/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-kube-api-access-c8qfw\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.807589 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-config\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.807630 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-dns-svc\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.810327 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-config\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.811357 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-dns-svc\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.840246 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8qfw\" (UniqueName: \"kubernetes.io/projected/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-kube-api-access-c8qfw\") pod \"dnsmasq-dns-7b46594d6f-72nrw\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.977746 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:02 crc kubenswrapper[4492]: I0220 06:54:02.993176 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5669d55c-5n8bd"] Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.048863 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b885f695-7kw2n"] Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.051030 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.072540 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b885f695-7kw2n"] Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.121134 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-config\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.121211 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-dns-svc\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.121257 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgvsf\" (UniqueName: \"kubernetes.io/projected/fdf66c08-5b42-418e-b48b-d2f32e6830f7-kube-api-access-qgvsf\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.224159 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgvsf\" (UniqueName: \"kubernetes.io/projected/fdf66c08-5b42-418e-b48b-d2f32e6830f7-kube-api-access-qgvsf\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.224408 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-config\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.224435 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-dns-svc\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.225337 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-dns-svc\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.225532 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-config\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.246458 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgvsf\" (UniqueName: \"kubernetes.io/projected/fdf66c08-5b42-418e-b48b-d2f32e6830f7-kube-api-access-qgvsf\") pod \"dnsmasq-dns-8b885f695-7kw2n\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.443131 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.656900 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b46594d6f-72nrw"] Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.841991 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.843715 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.845701 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.847170 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.847932 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.848175 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.848273 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.848344 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tnqs6" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.848434 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.872804 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" event={"ID":"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5","Type":"ContainerStarted","Data":"e7a2bc71d5c4e771f3f6e008ba317ca53258c008b4bc810e08921eff4ba1b885"} Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.875176 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 06:54:03 crc kubenswrapper[4492]: I0220 06:54:03.904047 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b885f695-7kw2n"] Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049007 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049077 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2719d90-a43e-4b5a-b4ad-e948f75c763a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049129 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049164 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2719d90-a43e-4b5a-b4ad-e948f75c763a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049262 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049288 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049325 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mvg6\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-kube-api-access-4mvg6\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049367 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-config-data\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049432 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049468 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.049552 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152055 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152118 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2719d90-a43e-4b5a-b4ad-e948f75c763a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152171 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152195 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2719d90-a43e-4b5a-b4ad-e948f75c763a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152229 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152249 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152299 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mvg6\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-kube-api-access-4mvg6\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152345 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-config-data\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152437 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152496 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152539 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.152931 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.153309 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.153647 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.155529 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-config-data\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.155970 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.159141 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.163889 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2719d90-a43e-4b5a-b4ad-e948f75c763a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.174687 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.176888 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.177460 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2719d90-a43e-4b5a-b4ad-e948f75c763a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.185444 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mvg6\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-kube-api-access-4mvg6\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.219809 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.244648 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.244838 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.253352 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.253610 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wtkz5" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.253795 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.254059 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.254317 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.254499 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.254884 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.283542 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357269 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357318 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357349 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c936d390-aaa6-4904-a1ca-357e2dac4cde-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357372 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c936d390-aaa6-4904-a1ca-357e2dac4cde-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357409 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357465 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357535 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357562 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357601 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpjhf\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-kube-api-access-hpjhf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357646 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.357695 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.459893 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c936d390-aaa6-4904-a1ca-357e2dac4cde-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.459976 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460050 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460124 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460173 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460222 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpjhf\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-kube-api-access-hpjhf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460271 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460344 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460398 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460432 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.460498 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c936d390-aaa6-4904-a1ca-357e2dac4cde-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.461655 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.465113 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.465911 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.465954 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.466303 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.466526 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.467704 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c936d390-aaa6-4904-a1ca-357e2dac4cde-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.470520 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.471958 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.472019 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c936d390-aaa6-4904-a1ca-357e2dac4cde-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.486934 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpjhf\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-kube-api-access-hpjhf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.489726 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.501445 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.595650 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:54:04 crc kubenswrapper[4492]: I0220 06:54:04.913943 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" event={"ID":"fdf66c08-5b42-418e-b48b-d2f32e6830f7","Type":"ContainerStarted","Data":"a1c0523204cf0a641ab49086f4f1d51344ef06ed84171b126ce7ab0de02cf0c0"} Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.066733 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 06:54:05 crc kubenswrapper[4492]: W0220 06:54:05.075832 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2719d90_a43e_4b5a_b4ad_e948f75c763a.slice/crio-aa76559b20f9369723626daba80900b006c6c6528401ac3da0d02869127ce79c WatchSource:0}: Error finding container aa76559b20f9369723626daba80900b006c6c6528401ac3da0d02869127ce79c: Status 404 returned error can't find the container with id aa76559b20f9369723626daba80900b006c6c6528401ac3da0d02869127ce79c Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.200820 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 06:54:05 crc kubenswrapper[4492]: W0220 06:54:05.215198 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc936d390_aaa6_4904_a1ca_357e2dac4cde.slice/crio-6f65f0257cb727f9b649ed38ca7978d001b557c6a531ae174240b3924719bf89 WatchSource:0}: Error finding container 6f65f0257cb727f9b649ed38ca7978d001b557c6a531ae174240b3924719bf89: Status 404 returned error can't find the container with id 6f65f0257cb727f9b649ed38ca7978d001b557c6a531ae174240b3924719bf89 Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.374551 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.376444 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.386324 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.387444 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.390027 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.411084 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.411976 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mqbjx" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.472636 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518530 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a40f3503-0190-40c1-bbfb-17a454347633-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518580 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-kolla-config\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518650 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k9j2\" (UniqueName: \"kubernetes.io/projected/a40f3503-0190-40c1-bbfb-17a454347633-kube-api-access-9k9j2\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518690 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-config-data-default\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518714 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518760 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a40f3503-0190-40c1-bbfb-17a454347633-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518794 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40f3503-0190-40c1-bbfb-17a454347633-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.518874 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.621991 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.622093 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a40f3503-0190-40c1-bbfb-17a454347633-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.622156 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-kolla-config\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.622237 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k9j2\" (UniqueName: \"kubernetes.io/projected/a40f3503-0190-40c1-bbfb-17a454347633-kube-api-access-9k9j2\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.622286 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-config-data-default\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.622307 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.622416 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a40f3503-0190-40c1-bbfb-17a454347633-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.622458 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40f3503-0190-40c1-bbfb-17a454347633-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.623162 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.623844 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-config-data-default\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.624122 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a40f3503-0190-40c1-bbfb-17a454347633-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.624559 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-kolla-config\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.626018 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40f3503-0190-40c1-bbfb-17a454347633-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.634317 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a40f3503-0190-40c1-bbfb-17a454347633-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.652339 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40f3503-0190-40c1-bbfb-17a454347633-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.670421 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k9j2\" (UniqueName: \"kubernetes.io/projected/a40f3503-0190-40c1-bbfb-17a454347633-kube-api-access-9k9j2\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.694681 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"a40f3503-0190-40c1-bbfb-17a454347633\") " pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.751082 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.938454 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c936d390-aaa6-4904-a1ca-357e2dac4cde","Type":"ContainerStarted","Data":"6f65f0257cb727f9b649ed38ca7978d001b557c6a531ae174240b3924719bf89"} Feb 20 06:54:05 crc kubenswrapper[4492]: I0220 06:54:05.944648 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a2719d90-a43e-4b5a-b4ad-e948f75c763a","Type":"ContainerStarted","Data":"aa76559b20f9369723626daba80900b006c6c6528401ac3da0d02869127ce79c"} Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.200962 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.869326 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.872399 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.882034 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.882404 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.883161 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.890901 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-qkd4l" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.899197 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.951449 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.951534 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.951564 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.951602 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.951638 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.951653 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.952043 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.952091 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5jqp\" (UniqueName: \"kubernetes.io/projected/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-kube-api-access-z5jqp\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.962559 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.963454 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.965237 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.965436 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-rmcp2" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.975894 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.990220 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 20 06:54:06 crc kubenswrapper[4492]: I0220 06:54:06.994730 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a40f3503-0190-40c1-bbfb-17a454347633","Type":"ContainerStarted","Data":"db1a4eb2277ffc039bb123f1d50a1c321ae95fc6e3fda49aae27c8d6d41fc58d"} Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054370 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054433 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/522bd757-e09f-424a-8459-57a7fb68952f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054462 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054546 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054584 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522bd757-e09f-424a-8459-57a7fb68952f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054606 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77wbk\" (UniqueName: \"kubernetes.io/projected/522bd757-e09f-424a-8459-57a7fb68952f-kube-api-access-77wbk\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054687 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.054801 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5jqp\" (UniqueName: \"kubernetes.io/projected/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-kube-api-access-z5jqp\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.055081 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.055143 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/522bd757-e09f-424a-8459-57a7fb68952f-config-data\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.055178 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.055203 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/522bd757-e09f-424a-8459-57a7fb68952f-kolla-config\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.055226 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.056055 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.056325 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.056319 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.056673 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.057852 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.082406 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.083281 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.090858 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5jqp\" (UniqueName: \"kubernetes.io/projected/4f0f24b3-67f0-4232-a358-0e1ae0b25f2f-kube-api-access-z5jqp\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.105516 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f\") " pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.157189 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/522bd757-e09f-424a-8459-57a7fb68952f-config-data\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.157241 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/522bd757-e09f-424a-8459-57a7fb68952f-kolla-config\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.157270 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/522bd757-e09f-424a-8459-57a7fb68952f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.157293 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522bd757-e09f-424a-8459-57a7fb68952f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.157310 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77wbk\" (UniqueName: \"kubernetes.io/projected/522bd757-e09f-424a-8459-57a7fb68952f-kube-api-access-77wbk\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.158337 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/522bd757-e09f-424a-8459-57a7fb68952f-kolla-config\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.158664 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/522bd757-e09f-424a-8459-57a7fb68952f-config-data\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.177931 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522bd757-e09f-424a-8459-57a7fb68952f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.182072 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/522bd757-e09f-424a-8459-57a7fb68952f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.206533 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77wbk\" (UniqueName: \"kubernetes.io/projected/522bd757-e09f-424a-8459-57a7fb68952f-kube-api-access-77wbk\") pod \"memcached-0\" (UID: \"522bd757-e09f-424a-8459-57a7fb68952f\") " pod="openstack/memcached-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.229921 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:07 crc kubenswrapper[4492]: I0220 06:54:07.291766 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 20 06:54:08 crc kubenswrapper[4492]: I0220 06:54:08.008800 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 20 06:54:08 crc kubenswrapper[4492]: W0220 06:54:08.047933 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f0f24b3_67f0_4232_a358_0e1ae0b25f2f.slice/crio-6cd5b60a22509aa91fc28dd7aa95928d63031f0ae795c15faabd281b6b819a52 WatchSource:0}: Error finding container 6cd5b60a22509aa91fc28dd7aa95928d63031f0ae795c15faabd281b6b819a52: Status 404 returned error can't find the container with id 6cd5b60a22509aa91fc28dd7aa95928d63031f0ae795c15faabd281b6b819a52 Feb 20 06:54:08 crc kubenswrapper[4492]: I0220 06:54:08.172329 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.050328 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"522bd757-e09f-424a-8459-57a7fb68952f","Type":"ContainerStarted","Data":"ba25e76aaae7569e86a5def38d71f8b36d9a6cd131fc3731a789f643e86d7762"} Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.083918 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f","Type":"ContainerStarted","Data":"6cd5b60a22509aa91fc28dd7aa95928d63031f0ae795c15faabd281b6b819a52"} Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.394071 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.398751 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.412033 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-xm49w" Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.421666 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.523262 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjncw\" (UniqueName: \"kubernetes.io/projected/03261a48-3a36-4223-b088-8d13030bdf81-kube-api-access-vjncw\") pod \"kube-state-metrics-0\" (UID: \"03261a48-3a36-4223-b088-8d13030bdf81\") " pod="openstack/kube-state-metrics-0" Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.625638 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjncw\" (UniqueName: \"kubernetes.io/projected/03261a48-3a36-4223-b088-8d13030bdf81-kube-api-access-vjncw\") pod \"kube-state-metrics-0\" (UID: \"03261a48-3a36-4223-b088-8d13030bdf81\") " pod="openstack/kube-state-metrics-0" Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.673585 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjncw\" (UniqueName: \"kubernetes.io/projected/03261a48-3a36-4223-b088-8d13030bdf81-kube-api-access-vjncw\") pod \"kube-state-metrics-0\" (UID: \"03261a48-3a36-4223-b088-8d13030bdf81\") " pod="openstack/kube-state-metrics-0" Feb 20 06:54:09 crc kubenswrapper[4492]: I0220 06:54:09.743099 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 06:54:10 crc kubenswrapper[4492]: I0220 06:54:10.464916 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:54:10 crc kubenswrapper[4492]: W0220 06:54:10.500521 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03261a48_3a36_4223_b088_8d13030bdf81.slice/crio-6bf7bb27cfa1bff20688e19b4fda54af6dab7e2fae9114ed754d9afbe11b4aca WatchSource:0}: Error finding container 6bf7bb27cfa1bff20688e19b4fda54af6dab7e2fae9114ed754d9afbe11b4aca: Status 404 returned error can't find the container with id 6bf7bb27cfa1bff20688e19b4fda54af6dab7e2fae9114ed754d9afbe11b4aca Feb 20 06:54:11 crc kubenswrapper[4492]: I0220 06:54:11.123580 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03261a48-3a36-4223-b088-8d13030bdf81","Type":"ContainerStarted","Data":"6bf7bb27cfa1bff20688e19b4fda54af6dab7e2fae9114ed754d9afbe11b4aca"} Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.634916 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6k7g9"] Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.636152 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.645702 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.645976 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-ggfhz" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.655949 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.669760 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-t7csh"] Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.671205 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.686508 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k7g9"] Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.715167 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t7csh"] Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.723020 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-scripts\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.723103 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-combined-ca-bundle\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.723293 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z689z\" (UniqueName: \"kubernetes.io/projected/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-kube-api-access-z689z\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.723318 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-ovn-controller-tls-certs\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.723339 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-log-ovn\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.723363 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-run-ovn\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.723379 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-run\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.824758 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ggql\" (UniqueName: \"kubernetes.io/projected/3778ef17-0366-4509-8b07-64c301e9de60-kube-api-access-6ggql\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.824813 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3778ef17-0366-4509-8b07-64c301e9de60-scripts\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.824882 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-run\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.824909 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z689z\" (UniqueName: \"kubernetes.io/projected/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-kube-api-access-z689z\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.824986 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-ovn-controller-tls-certs\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825034 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-log-ovn\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825078 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-run-ovn\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825098 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-run\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825148 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-scripts\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825170 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-lib\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825242 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-combined-ca-bundle\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825279 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-etc-ovs\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825322 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-log\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.825830 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-run\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.827263 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-scripts\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.827709 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-log-ovn\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.828053 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-var-run-ovn\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.861728 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-ovn-controller-tls-certs\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.877231 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z689z\" (UniqueName: \"kubernetes.io/projected/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-kube-api-access-z689z\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.899641 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb-combined-ca-bundle\") pod \"ovn-controller-6k7g9\" (UID: \"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb\") " pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.927139 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ggql\" (UniqueName: \"kubernetes.io/projected/3778ef17-0366-4509-8b07-64c301e9de60-kube-api-access-6ggql\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.927183 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3778ef17-0366-4509-8b07-64c301e9de60-scripts\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.927226 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-run\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.927276 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-lib\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.927332 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-etc-ovs\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.927362 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-log\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.927548 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-log\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.928772 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-lib\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.928869 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-var-run\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.929014 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3778ef17-0366-4509-8b07-64c301e9de60-etc-ovs\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.929536 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3778ef17-0366-4509-8b07-64c301e9de60-scripts\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.955954 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ggql\" (UniqueName: \"kubernetes.io/projected/3778ef17-0366-4509-8b07-64c301e9de60-kube-api-access-6ggql\") pod \"ovn-controller-ovs-t7csh\" (UID: \"3778ef17-0366-4509-8b07-64c301e9de60\") " pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:12 crc kubenswrapper[4492]: I0220 06:54:12.980581 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.008747 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.335893 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.341666 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.344746 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.345274 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-hn7cd" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.345376 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.346743 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.349149 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.349311 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.436773 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.436817 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.436850 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.436881 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8d9698-66bb-44b6-b225-e77fc088ff02-config\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.436916 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/3b8d9698-66bb-44b6-b225-e77fc088ff02-kube-api-access-2gj25\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.436940 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8d9698-66bb-44b6-b225-e77fc088ff02-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.437002 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.437027 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3b8d9698-66bb-44b6-b225-e77fc088ff02-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539565 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539629 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539682 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539715 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8d9698-66bb-44b6-b225-e77fc088ff02-config\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539761 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/3b8d9698-66bb-44b6-b225-e77fc088ff02-kube-api-access-2gj25\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539786 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8d9698-66bb-44b6-b225-e77fc088ff02-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539858 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.539886 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3b8d9698-66bb-44b6-b225-e77fc088ff02-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.540558 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3b8d9698-66bb-44b6-b225-e77fc088ff02-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.541317 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.542610 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b8d9698-66bb-44b6-b225-e77fc088ff02-config\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.542737 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3b8d9698-66bb-44b6-b225-e77fc088ff02-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.544116 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.548592 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.554054 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d9698-66bb-44b6-b225-e77fc088ff02-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.563844 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gj25\" (UniqueName: \"kubernetes.io/projected/3b8d9698-66bb-44b6-b225-e77fc088ff02-kube-api-access-2gj25\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.567163 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3b8d9698-66bb-44b6-b225-e77fc088ff02\") " pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:13 crc kubenswrapper[4492]: I0220 06:54:13.659720 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.428033 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.429800 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.433660 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lvtqb" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.434171 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.435819 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.435999 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.473140 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.550976 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.551049 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbdd3644-632d-4bf1-90d0-0fb9a3392884-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.551076 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.551132 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww7tk\" (UniqueName: \"kubernetes.io/projected/dbdd3644-632d-4bf1-90d0-0fb9a3392884-kube-api-access-ww7tk\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.551296 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbdd3644-632d-4bf1-90d0-0fb9a3392884-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.551439 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbdd3644-632d-4bf1-90d0-0fb9a3392884-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.551518 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.551546 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653027 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbdd3644-632d-4bf1-90d0-0fb9a3392884-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653090 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbdd3644-632d-4bf1-90d0-0fb9a3392884-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653129 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653162 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653198 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653230 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbdd3644-632d-4bf1-90d0-0fb9a3392884-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653246 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.653279 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww7tk\" (UniqueName: \"kubernetes.io/projected/dbdd3644-632d-4bf1-90d0-0fb9a3392884-kube-api-access-ww7tk\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.654343 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.654491 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dbdd3644-632d-4bf1-90d0-0fb9a3392884-config\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.655193 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbdd3644-632d-4bf1-90d0-0fb9a3392884-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.655750 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dbdd3644-632d-4bf1-90d0-0fb9a3392884-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.667566 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.668048 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.671322 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww7tk\" (UniqueName: \"kubernetes.io/projected/dbdd3644-632d-4bf1-90d0-0fb9a3392884-kube-api-access-ww7tk\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.674540 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbdd3644-632d-4bf1-90d0-0fb9a3392884-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.686269 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"dbdd3644-632d-4bf1-90d0-0fb9a3392884\") " pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:16 crc kubenswrapper[4492]: I0220 06:54:16.755436 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:23 crc kubenswrapper[4492]: I0220 06:54:23.042424 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t7csh"] Feb 20 06:54:23 crc kubenswrapper[4492]: W0220 06:54:23.162403 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3778ef17_0366_4509_8b07_64c301e9de60.slice/crio-9f41dd1aba9466c5a92b88c6bb0e58894e582f446a43373e900e7433b4e621d4 WatchSource:0}: Error finding container 9f41dd1aba9466c5a92b88c6bb0e58894e582f446a43373e900e7433b4e621d4: Status 404 returned error can't find the container with id 9f41dd1aba9466c5a92b88c6bb0e58894e582f446a43373e900e7433b4e621d4 Feb 20 06:54:23 crc kubenswrapper[4492]: I0220 06:54:23.167772 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 06:54:23 crc kubenswrapper[4492]: I0220 06:54:23.316366 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7csh" event={"ID":"3778ef17-0366-4509-8b07-64c301e9de60","Type":"ContainerStarted","Data":"9f41dd1aba9466c5a92b88c6bb0e58894e582f446a43373e900e7433b4e621d4"} Feb 20 06:54:23 crc kubenswrapper[4492]: I0220 06:54:23.438360 4492 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 20 06:54:23 crc kubenswrapper[4492]: I0220 06:54:23.677157 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k7g9"] Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.853852 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-zwhpv"] Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.856047 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.859835 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.860898 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d4345dc0-8026-48f6-ab0f-57d54ae69433-ovs-rundir\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.861008 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d4345dc0-8026-48f6-ab0f-57d54ae69433-ovn-rundir\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.861128 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4345dc0-8026-48f6-ab0f-57d54ae69433-combined-ca-bundle\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.861151 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4345dc0-8026-48f6-ab0f-57d54ae69433-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.861266 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345dc0-8026-48f6-ab0f-57d54ae69433-config\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.861294 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjcpv\" (UniqueName: \"kubernetes.io/projected/d4345dc0-8026-48f6-ab0f-57d54ae69433-kube-api-access-vjcpv\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.866933 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zwhpv"] Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.962756 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d4345dc0-8026-48f6-ab0f-57d54ae69433-ovn-rundir\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.963605 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4345dc0-8026-48f6-ab0f-57d54ae69433-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.963630 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4345dc0-8026-48f6-ab0f-57d54ae69433-combined-ca-bundle\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.963709 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345dc0-8026-48f6-ab0f-57d54ae69433-config\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.963758 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjcpv\" (UniqueName: \"kubernetes.io/projected/d4345dc0-8026-48f6-ab0f-57d54ae69433-kube-api-access-vjcpv\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.963830 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d4345dc0-8026-48f6-ab0f-57d54ae69433-ovs-rundir\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.964003 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/d4345dc0-8026-48f6-ab0f-57d54ae69433-ovs-rundir\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.964125 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/d4345dc0-8026-48f6-ab0f-57d54ae69433-ovn-rundir\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.965130 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4345dc0-8026-48f6-ab0f-57d54ae69433-config\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.978146 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4345dc0-8026-48f6-ab0f-57d54ae69433-combined-ca-bundle\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.982596 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjcpv\" (UniqueName: \"kubernetes.io/projected/d4345dc0-8026-48f6-ab0f-57d54ae69433-kube-api-access-vjcpv\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:25 crc kubenswrapper[4492]: I0220 06:54:25.983144 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4345dc0-8026-48f6-ab0f-57d54ae69433-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zwhpv\" (UID: \"d4345dc0-8026-48f6-ab0f-57d54ae69433\") " pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:26 crc kubenswrapper[4492]: I0220 06:54:26.185741 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zwhpv" Feb 20 06:54:31 crc kubenswrapper[4492]: W0220 06:54:31.189670 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53c1e7c7_1b2d_4e23_b4bf_fca7fe5833eb.slice/crio-6c59da0864990d85cff6ad7615f01e8800e66903e03d8157641388d56b1cc0fe WatchSource:0}: Error finding container 6c59da0864990d85cff6ad7615f01e8800e66903e03d8157641388d56b1cc0fe: Status 404 returned error can't find the container with id 6c59da0864990d85cff6ad7615f01e8800e66903e03d8157641388d56b1cc0fe Feb 20 06:54:31 crc kubenswrapper[4492]: I0220 06:54:31.429510 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k7g9" event={"ID":"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb","Type":"ContainerStarted","Data":"6c59da0864990d85cff6ad7615f01e8800e66903e03d8157641388d56b1cc0fe"} Feb 20 06:54:31 crc kubenswrapper[4492]: I0220 06:54:31.698023 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 20 06:54:32 crc kubenswrapper[4492]: E0220 06:54:32.941744 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923" Feb 20 06:54:32 crc kubenswrapper[4492]: E0220 06:54:32.941796 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923" Feb 20 06:54:32 crc kubenswrapper[4492]: E0220 06:54:32.941919 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hpjhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(c936d390-aaa6-4904-a1ca-357e2dac4cde): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:54:32 crc kubenswrapper[4492]: E0220 06:54:32.943129 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" Feb 20 06:54:33 crc kubenswrapper[4492]: E0220 06:54:33.447217 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" Feb 20 06:54:34 crc kubenswrapper[4492]: W0220 06:54:34.989883 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8d9698_66bb_44b6_b225_e77fc088ff02.slice/crio-905558f7deb866968e0635f3e245e66dd6c47047bc13fe0eb3d67f1e91d88eb1 WatchSource:0}: Error finding container 905558f7deb866968e0635f3e245e66dd6c47047bc13fe0eb3d67f1e91d88eb1: Status 404 returned error can't find the container with id 905558f7deb866968e0635f3e245e66dd6c47047bc13fe0eb3d67f1e91d88eb1 Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.011778 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:f0473f3e18dd17d7021c02e991298923" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.011841 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:f0473f3e18dd17d7021c02e991298923" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.011992 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:f0473f3e18dd17d7021c02e991298923,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9k9j2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(a40f3503-0190-40c1-bbfb-17a454347633): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.013252 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="a40f3503-0190-40c1-bbfb-17a454347633" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.028024 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.028098 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.028276 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4mvg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(a2719d90-a43e-4b5a-b4ad-e948f75c763a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.029440 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" Feb 20 06:54:35 crc kubenswrapper[4492]: I0220 06:54:35.480024 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3b8d9698-66bb-44b6-b225-e77fc088ff02","Type":"ContainerStarted","Data":"905558f7deb866968e0635f3e245e66dd6c47047bc13fe0eb3d67f1e91d88eb1"} Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.481107 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:f0473f3e18dd17d7021c02e991298923\\\"\"" pod="openstack/openstack-galera-0" podUID="a40f3503-0190-40c1-bbfb-17a454347633" Feb 20 06:54:35 crc kubenswrapper[4492]: E0220 06:54:35.481361 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:f0473f3e18dd17d7021c02e991298923\\\"\"" pod="openstack/rabbitmq-server-0" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" Feb 20 06:54:35 crc kubenswrapper[4492]: I0220 06:54:35.527345 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 20 06:54:36 crc kubenswrapper[4492]: W0220 06:54:36.212395 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4345dc0_8026_48f6_ab0f_57d54ae69433.slice/crio-3ac272c00957a0f81a5a2bc41fec15c0cb7ee2318dc7af325a4a3eb08326756d WatchSource:0}: Error finding container 3ac272c00957a0f81a5a2bc41fec15c0cb7ee2318dc7af325a4a3eb08326756d: Status 404 returned error can't find the container with id 3ac272c00957a0f81a5a2bc41fec15c0cb7ee2318dc7af325a4a3eb08326756d Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.218549 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zwhpv"] Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.489194 4492 generic.go:334] "Generic (PLEG): container finished" podID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerID="c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc" exitCode=0 Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.489285 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" event={"ID":"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5","Type":"ContainerDied","Data":"c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.492076 4492 generic.go:334] "Generic (PLEG): container finished" podID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerID="c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926" exitCode=0 Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.492146 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" event={"ID":"fdf66c08-5b42-418e-b48b-d2f32e6830f7","Type":"ContainerDied","Data":"c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.495245 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbdd3644-632d-4bf1-90d0-0fb9a3392884","Type":"ContainerStarted","Data":"621b9d58eb9be59a9bdea975a94373315192322d566b6c29b532a202ac1572ec"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.498688 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zwhpv" event={"ID":"d4345dc0-8026-48f6-ab0f-57d54ae69433","Type":"ContainerStarted","Data":"3ac272c00957a0f81a5a2bc41fec15c0cb7ee2318dc7af325a4a3eb08326756d"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.501187 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f","Type":"ContainerStarted","Data":"1525f1b4a65d6f8939d237cac094c1b5d7c8d746d7ea6f226cd638228e985e4e"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.515958 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03261a48-3a36-4223-b088-8d13030bdf81","Type":"ContainerStarted","Data":"d490c95ea1a9fe69fa81243207288d99ce3ad103fb59be07ef869a15d71816b7"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.516649 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.525388 4492 generic.go:334] "Generic (PLEG): container finished" podID="9d7bd976-285d-4222-b9c7-37c27bd6fd8a" containerID="502cf203ff41993eefacef007c40c15e9602075e06053dcc076542f0f29e5567" exitCode=0 Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.525467 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" event={"ID":"9d7bd976-285d-4222-b9c7-37c27bd6fd8a","Type":"ContainerDied","Data":"502cf203ff41993eefacef007c40c15e9602075e06053dcc076542f0f29e5567"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.529750 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"522bd757-e09f-424a-8459-57a7fb68952f","Type":"ContainerStarted","Data":"e4b547577471108b857806a72bd8ef639dc1e2641c5debddd487f9a1b4e4b9f6"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.530143 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.534914 4492 generic.go:334] "Generic (PLEG): container finished" podID="c02da0cb-f83e-468b-b8e0-2bfc7126da7f" containerID="9c4f390da6e249dd2e75e282408c92e2eb1a44b24215db8b9b5a17e2b0980df3" exitCode=0 Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.535053 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" event={"ID":"c02da0cb-f83e-468b-b8e0-2bfc7126da7f","Type":"ContainerDied","Data":"9c4f390da6e249dd2e75e282408c92e2eb1a44b24215db8b9b5a17e2b0980df3"} Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.630959 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.875837648 podStartE2EDuration="30.630938257s" podCreationTimestamp="2026-02-20 06:54:06 +0000 UTC" firstStartedPulling="2026-02-20 06:54:08.219495117 +0000 UTC m=+804.990784095" lastFinishedPulling="2026-02-20 06:54:35.974595726 +0000 UTC m=+832.745884704" observedRunningTime="2026-02-20 06:54:36.600554086 +0000 UTC m=+833.371843074" watchObservedRunningTime="2026-02-20 06:54:36.630938257 +0000 UTC m=+833.402227235" Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.659033 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.004711173 podStartE2EDuration="27.659014197s" podCreationTimestamp="2026-02-20 06:54:09 +0000 UTC" firstStartedPulling="2026-02-20 06:54:10.506405394 +0000 UTC m=+807.277694371" lastFinishedPulling="2026-02-20 06:54:23.160708417 +0000 UTC m=+819.931997395" observedRunningTime="2026-02-20 06:54:36.633750346 +0000 UTC m=+833.405039324" watchObservedRunningTime="2026-02-20 06:54:36.659014197 +0000 UTC m=+833.430303175" Feb 20 06:54:36 crc kubenswrapper[4492]: E0220 06:54:36.768466 4492 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Feb 20 06:54:36 crc kubenswrapper[4492]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 20 06:54:36 crc kubenswrapper[4492]: > podSandboxID="e7a2bc71d5c4e771f3f6e008ba317ca53258c008b4bc810e08921eff4ba1b885" Feb 20 06:54:36 crc kubenswrapper[4492]: E0220 06:54:36.769123 4492 kuberuntime_manager.go:1274] "Unhandled Error" err=< Feb 20 06:54:36 crc kubenswrapper[4492]: container &Container{Name:dnsmasq-dns,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c8qfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7b46594d6f-72nrw_openstack(a88f30d1-ceb3-423a-bcc9-ef66643cd6d5): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Feb 20 06:54:36 crc kubenswrapper[4492]: > logger="UnhandledError" Feb 20 06:54:36 crc kubenswrapper[4492]: E0220 06:54:36.771027 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.847824 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:36 crc kubenswrapper[4492]: I0220 06:54:36.972204 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.005260 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpl9v\" (UniqueName: \"kubernetes.io/projected/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-kube-api-access-tpl9v\") pod \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.005354 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-config\") pod \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.005385 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-dns-svc\") pod \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\" (UID: \"9d7bd976-285d-4222-b9c7-37c27bd6fd8a\") " Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.015113 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-kube-api-access-tpl9v" (OuterVolumeSpecName: "kube-api-access-tpl9v") pod "9d7bd976-285d-4222-b9c7-37c27bd6fd8a" (UID: "9d7bd976-285d-4222-b9c7-37c27bd6fd8a"). InnerVolumeSpecName "kube-api-access-tpl9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.027181 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-config" (OuterVolumeSpecName: "config") pod "9d7bd976-285d-4222-b9c7-37c27bd6fd8a" (UID: "9d7bd976-285d-4222-b9c7-37c27bd6fd8a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.035812 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d7bd976-285d-4222-b9c7-37c27bd6fd8a" (UID: "9d7bd976-285d-4222-b9c7-37c27bd6fd8a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.107068 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-config\") pod \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.107147 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts7q2\" (UniqueName: \"kubernetes.io/projected/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-kube-api-access-ts7q2\") pod \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\" (UID: \"c02da0cb-f83e-468b-b8e0-2bfc7126da7f\") " Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.107410 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpl9v\" (UniqueName: \"kubernetes.io/projected/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-kube-api-access-tpl9v\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.107609 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.107633 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d7bd976-285d-4222-b9c7-37c27bd6fd8a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.122227 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-kube-api-access-ts7q2" (OuterVolumeSpecName: "kube-api-access-ts7q2") pod "c02da0cb-f83e-468b-b8e0-2bfc7126da7f" (UID: "c02da0cb-f83e-468b-b8e0-2bfc7126da7f"). InnerVolumeSpecName "kube-api-access-ts7q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.124250 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-config" (OuterVolumeSpecName: "config") pod "c02da0cb-f83e-468b-b8e0-2bfc7126da7f" (UID: "c02da0cb-f83e-468b-b8e0-2bfc7126da7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.211701 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.211734 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts7q2\" (UniqueName: \"kubernetes.io/projected/c02da0cb-f83e-468b-b8e0-2bfc7126da7f-kube-api-access-ts7q2\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.551262 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" event={"ID":"c02da0cb-f83e-468b-b8e0-2bfc7126da7f","Type":"ContainerDied","Data":"8ff24cf66b682f30cb30467f5fea7864cb43b6a018f429b156e49c3e06823972"} Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.551281 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5669d55c-5n8bd" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.556621 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.558119 4492 scope.go:117] "RemoveContainer" containerID="9c4f390da6e249dd2e75e282408c92e2eb1a44b24215db8b9b5a17e2b0980df3" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.560083 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.578320 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" podStartSLOduration=2.572131559 podStartE2EDuration="34.578280641s" podCreationTimestamp="2026-02-20 06:54:03 +0000 UTC" firstStartedPulling="2026-02-20 06:54:03.92553796 +0000 UTC m=+800.696826939" lastFinishedPulling="2026-02-20 06:54:35.931687044 +0000 UTC m=+832.702976021" observedRunningTime="2026-02-20 06:54:37.578050257 +0000 UTC m=+834.349339234" watchObservedRunningTime="2026-02-20 06:54:37.578280641 +0000 UTC m=+834.349569619" Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.585025 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" event={"ID":"fdf66c08-5b42-418e-b48b-d2f32e6830f7","Type":"ContainerStarted","Data":"ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f"} Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.585064 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c87b565c-qrkbl" event={"ID":"9d7bd976-285d-4222-b9c7-37c27bd6fd8a","Type":"ContainerDied","Data":"22ba3663934c47860afc326655811585b5a4f829bbf9685b6c6254a94fcd05cc"} Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.639258 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68c87b565c-qrkbl"] Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.652762 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68c87b565c-qrkbl"] Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.682658 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5669d55c-5n8bd"] Feb 20 06:54:37 crc kubenswrapper[4492]: I0220 06:54:37.688031 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c5669d55c-5n8bd"] Feb 20 06:54:38 crc kubenswrapper[4492]: I0220 06:54:38.242653 4492 scope.go:117] "RemoveContainer" containerID="502cf203ff41993eefacef007c40c15e9602075e06053dcc076542f0f29e5567" Feb 20 06:54:38 crc kubenswrapper[4492]: I0220 06:54:38.573288 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zwhpv" event={"ID":"d4345dc0-8026-48f6-ab0f-57d54ae69433","Type":"ContainerStarted","Data":"a47afc26df309bf0536992d6d055c2dd6a8c054f351c92c7486c68f27b06d192"} Feb 20 06:54:38 crc kubenswrapper[4492]: I0220 06:54:38.579489 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" event={"ID":"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5","Type":"ContainerStarted","Data":"7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c"} Feb 20 06:54:38 crc kubenswrapper[4492]: I0220 06:54:38.579931 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:38 crc kubenswrapper[4492]: I0220 06:54:38.598192 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-zwhpv" podStartSLOduration=11.519678622 podStartE2EDuration="13.598178038s" podCreationTimestamp="2026-02-20 06:54:25 +0000 UTC" firstStartedPulling="2026-02-20 06:54:36.216512711 +0000 UTC m=+832.987801689" lastFinishedPulling="2026-02-20 06:54:38.295012137 +0000 UTC m=+835.066301105" observedRunningTime="2026-02-20 06:54:38.596023458 +0000 UTC m=+835.367312436" watchObservedRunningTime="2026-02-20 06:54:38.598178038 +0000 UTC m=+835.369467016" Feb 20 06:54:38 crc kubenswrapper[4492]: I0220 06:54:38.625008 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" podStartSLOduration=4.322966563 podStartE2EDuration="36.624991088s" podCreationTimestamp="2026-02-20 06:54:02 +0000 UTC" firstStartedPulling="2026-02-20 06:54:03.675250912 +0000 UTC m=+800.446539890" lastFinishedPulling="2026-02-20 06:54:35.977275437 +0000 UTC m=+832.748564415" observedRunningTime="2026-02-20 06:54:38.620835267 +0000 UTC m=+835.392124245" watchObservedRunningTime="2026-02-20 06:54:38.624991088 +0000 UTC m=+835.396280067" Feb 20 06:54:38 crc kubenswrapper[4492]: I0220 06:54:38.952098 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b46594d6f-72nrw"] Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.003913 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b55c8b857-7s4rc"] Feb 20 06:54:39 crc kubenswrapper[4492]: E0220 06:54:39.004762 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c02da0cb-f83e-468b-b8e0-2bfc7126da7f" containerName="init" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.004784 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c02da0cb-f83e-468b-b8e0-2bfc7126da7f" containerName="init" Feb 20 06:54:39 crc kubenswrapper[4492]: E0220 06:54:39.004829 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d7bd976-285d-4222-b9c7-37c27bd6fd8a" containerName="init" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.004836 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d7bd976-285d-4222-b9c7-37c27bd6fd8a" containerName="init" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.005215 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c02da0cb-f83e-468b-b8e0-2bfc7126da7f" containerName="init" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.005257 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d7bd976-285d-4222-b9c7-37c27bd6fd8a" containerName="init" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.018571 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.022206 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.046260 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b55c8b857-7s4rc"] Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.155718 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-dns-svc\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.155813 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-226ns\" (UniqueName: \"kubernetes.io/projected/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-kube-api-access-226ns\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.155881 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.155913 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-config\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.174894 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b885f695-7kw2n"] Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.207839 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59b55bff5-dnq8f"] Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.213888 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.215392 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b55bff5-dnq8f"] Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.216336 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.258268 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.258328 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-config\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.258460 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-dns-svc\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.258539 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-226ns\" (UniqueName: \"kubernetes.io/projected/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-kube-api-access-226ns\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.259316 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.259527 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-config\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.259571 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-dns-svc\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.297284 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-226ns\" (UniqueName: \"kubernetes.io/projected/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-kube-api-access-226ns\") pod \"dnsmasq-dns-b55c8b857-7s4rc\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.359838 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-sb\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.359904 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z22x5\" (UniqueName: \"kubernetes.io/projected/2558c7b3-5ab5-45d8-b183-4f74014349da-kube-api-access-z22x5\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.360023 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-config\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.360044 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-dns-svc\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.360064 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-nb\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.366520 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.461511 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-config\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.461568 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-dns-svc\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.461600 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-nb\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.461663 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-sb\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.461707 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z22x5\" (UniqueName: \"kubernetes.io/projected/2558c7b3-5ab5-45d8-b183-4f74014349da-kube-api-access-z22x5\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.463209 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-config\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.463891 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-sb\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.463917 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-nb\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.464561 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-dns-svc\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.481196 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z22x5\" (UniqueName: \"kubernetes.io/projected/2558c7b3-5ab5-45d8-b183-4f74014349da-kube-api-access-z22x5\") pod \"dnsmasq-dns-59b55bff5-dnq8f\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.578946 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d7bd976-285d-4222-b9c7-37c27bd6fd8a" path="/var/lib/kubelet/pods/9d7bd976-285d-4222-b9c7-37c27bd6fd8a/volumes" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.579615 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c02da0cb-f83e-468b-b8e0-2bfc7126da7f" path="/var/lib/kubelet/pods/c02da0cb-f83e-468b-b8e0-2bfc7126da7f/volumes" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.580113 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.594239 4492 generic.go:334] "Generic (PLEG): container finished" podID="4f0f24b3-67f0-4232-a358-0e1ae0b25f2f" containerID="1525f1b4a65d6f8939d237cac094c1b5d7c8d746d7ea6f226cd638228e985e4e" exitCode=0 Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.594386 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f","Type":"ContainerDied","Data":"1525f1b4a65d6f8939d237cac094c1b5d7c8d746d7ea6f226cd638228e985e4e"} Feb 20 06:54:39 crc kubenswrapper[4492]: I0220 06:54:39.854222 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b55c8b857-7s4rc"] Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.033875 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b55bff5-dnq8f"] Feb 20 06:54:40 crc kubenswrapper[4492]: W0220 06:54:40.042826 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2558c7b3_5ab5_45d8_b183_4f74014349da.slice/crio-ff0e92af632f294da9441252bac37e2c493522c59179ae30c825742cd8056ff9 WatchSource:0}: Error finding container ff0e92af632f294da9441252bac37e2c493522c59179ae30c825742cd8056ff9: Status 404 returned error can't find the container with id ff0e92af632f294da9441252bac37e2c493522c59179ae30c825742cd8056ff9 Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.616757 4492 generic.go:334] "Generic (PLEG): container finished" podID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerID="bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb" exitCode=0 Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.616809 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" event={"ID":"11ee4e73-dcc1-42d5-9b1f-c1a55935168c","Type":"ContainerDied","Data":"bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb"} Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.616859 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" event={"ID":"11ee4e73-dcc1-42d5-9b1f-c1a55935168c","Type":"ContainerStarted","Data":"7e6015c5295577b98bc629b3034ebccc89b8e9f9635289add1aa27280359b3e3"} Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.620339 4492 generic.go:334] "Generic (PLEG): container finished" podID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerID="8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92" exitCode=0 Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.620420 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" event={"ID":"2558c7b3-5ab5-45d8-b183-4f74014349da","Type":"ContainerDied","Data":"8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92"} Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.620451 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" event={"ID":"2558c7b3-5ab5-45d8-b183-4f74014349da","Type":"ContainerStarted","Data":"ff0e92af632f294da9441252bac37e2c493522c59179ae30c825742cd8056ff9"} Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.634764 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerName="dnsmasq-dns" containerID="cri-o://7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c" gracePeriod=10 Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.635156 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4f0f24b3-67f0-4232-a358-0e1ae0b25f2f","Type":"ContainerStarted","Data":"83db144b3f5503ad268bf8c0026cae617f76d2acc3a9cb83fff68c3c13856a45"} Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.637433 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" podUID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerName="dnsmasq-dns" containerID="cri-o://ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f" gracePeriod=10 Feb 20 06:54:40 crc kubenswrapper[4492]: I0220 06:54:40.672310 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.781313881 podStartE2EDuration="35.67229258s" podCreationTimestamp="2026-02-20 06:54:05 +0000 UTC" firstStartedPulling="2026-02-20 06:54:08.083606548 +0000 UTC m=+804.854895525" lastFinishedPulling="2026-02-20 06:54:35.974585247 +0000 UTC m=+832.745874224" observedRunningTime="2026-02-20 06:54:40.667754659 +0000 UTC m=+837.439043637" watchObservedRunningTime="2026-02-20 06:54:40.67229258 +0000 UTC m=+837.443581558" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.010199 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.105792 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-dns-svc\") pod \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.105879 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgvsf\" (UniqueName: \"kubernetes.io/projected/fdf66c08-5b42-418e-b48b-d2f32e6830f7-kube-api-access-qgvsf\") pod \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.106034 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-config\") pod \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\" (UID: \"fdf66c08-5b42-418e-b48b-d2f32e6830f7\") " Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.119664 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf66c08-5b42-418e-b48b-d2f32e6830f7-kube-api-access-qgvsf" (OuterVolumeSpecName: "kube-api-access-qgvsf") pod "fdf66c08-5b42-418e-b48b-d2f32e6830f7" (UID: "fdf66c08-5b42-418e-b48b-d2f32e6830f7"). InnerVolumeSpecName "kube-api-access-qgvsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.146355 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-config" (OuterVolumeSpecName: "config") pod "fdf66c08-5b42-418e-b48b-d2f32e6830f7" (UID: "fdf66c08-5b42-418e-b48b-d2f32e6830f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.152575 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fdf66c08-5b42-418e-b48b-d2f32e6830f7" (UID: "fdf66c08-5b42-418e-b48b-d2f32e6830f7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.172442 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.207768 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-config\") pod \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.207928 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8qfw\" (UniqueName: \"kubernetes.io/projected/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-kube-api-access-c8qfw\") pod \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.207966 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-dns-svc\") pod \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\" (UID: \"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5\") " Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.208330 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.208348 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fdf66c08-5b42-418e-b48b-d2f32e6830f7-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.208359 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgvsf\" (UniqueName: \"kubernetes.io/projected/fdf66c08-5b42-418e-b48b-d2f32e6830f7-kube-api-access-qgvsf\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.212030 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-kube-api-access-c8qfw" (OuterVolumeSpecName: "kube-api-access-c8qfw") pod "a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" (UID: "a88f30d1-ceb3-423a-bcc9-ef66643cd6d5"). InnerVolumeSpecName "kube-api-access-c8qfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.237460 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" (UID: "a88f30d1-ceb3-423a-bcc9-ef66643cd6d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.239657 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-config" (OuterVolumeSpecName: "config") pod "a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" (UID: "a88f30d1-ceb3-423a-bcc9-ef66643cd6d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.309626 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8qfw\" (UniqueName: \"kubernetes.io/projected/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-kube-api-access-c8qfw\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.309654 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.309664 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.647256 4492 generic.go:334] "Generic (PLEG): container finished" podID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerID="7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c" exitCode=0 Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.647382 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" event={"ID":"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5","Type":"ContainerDied","Data":"7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c"} Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.647994 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" event={"ID":"a88f30d1-ceb3-423a-bcc9-ef66643cd6d5","Type":"ContainerDied","Data":"e7a2bc71d5c4e771f3f6e008ba317ca53258c008b4bc810e08921eff4ba1b885"} Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.647401 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b46594d6f-72nrw" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.648064 4492 scope.go:117] "RemoveContainer" containerID="7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.653707 4492 generic.go:334] "Generic (PLEG): container finished" podID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerID="ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f" exitCode=0 Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.653851 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" event={"ID":"fdf66c08-5b42-418e-b48b-d2f32e6830f7","Type":"ContainerDied","Data":"ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f"} Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.653933 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" event={"ID":"fdf66c08-5b42-418e-b48b-d2f32e6830f7","Type":"ContainerDied","Data":"a1c0523204cf0a641ab49086f4f1d51344ef06ed84171b126ce7ab0de02cf0c0"} Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.653977 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b885f695-7kw2n" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.658093 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" event={"ID":"11ee4e73-dcc1-42d5-9b1f-c1a55935168c","Type":"ContainerStarted","Data":"1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac"} Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.658282 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.671357 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" event={"ID":"2558c7b3-5ab5-45d8-b183-4f74014349da","Type":"ContainerStarted","Data":"5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee"} Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.672361 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.681568 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b46594d6f-72nrw"] Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.693194 4492 scope.go:117] "RemoveContainer" containerID="c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.695737 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b46594d6f-72nrw"] Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.700311 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" podStartSLOduration=2.7002958230000003 podStartE2EDuration="2.700295823s" podCreationTimestamp="2026-02-20 06:54:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:54:41.687845742 +0000 UTC m=+838.459134709" watchObservedRunningTime="2026-02-20 06:54:41.700295823 +0000 UTC m=+838.471584802" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.732158 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" podStartSLOduration=3.732129116 podStartE2EDuration="3.732129116s" podCreationTimestamp="2026-02-20 06:54:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:54:41.710526093 +0000 UTC m=+838.481815072" watchObservedRunningTime="2026-02-20 06:54:41.732129116 +0000 UTC m=+838.503418083" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.732242 4492 scope.go:117] "RemoveContainer" containerID="7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c" Feb 20 06:54:41 crc kubenswrapper[4492]: E0220 06:54:41.732907 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c\": container with ID starting with 7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c not found: ID does not exist" containerID="7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.732977 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c"} err="failed to get container status \"7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c\": rpc error: code = NotFound desc = could not find container \"7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c\": container with ID starting with 7d1f9897e1fea025db7ef8b46792ca442bac5aa00a970eae51938ce3db477c0c not found: ID does not exist" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.733017 4492 scope.go:117] "RemoveContainer" containerID="c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc" Feb 20 06:54:41 crc kubenswrapper[4492]: E0220 06:54:41.733498 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc\": container with ID starting with c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc not found: ID does not exist" containerID="c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.733522 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc"} err="failed to get container status \"c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc\": rpc error: code = NotFound desc = could not find container \"c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc\": container with ID starting with c23418b10a6d28cc53ab3ce6c99f321b5589a33263b60f5f85730b4026d4d5dc not found: ID does not exist" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.733539 4492 scope.go:117] "RemoveContainer" containerID="ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.742120 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b885f695-7kw2n"] Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.747662 4492 scope.go:117] "RemoveContainer" containerID="c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.749640 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b885f695-7kw2n"] Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.763860 4492 scope.go:117] "RemoveContainer" containerID="ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f" Feb 20 06:54:41 crc kubenswrapper[4492]: E0220 06:54:41.764692 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f\": container with ID starting with ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f not found: ID does not exist" containerID="ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.764740 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f"} err="failed to get container status \"ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f\": rpc error: code = NotFound desc = could not find container \"ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f\": container with ID starting with ff4092b520a694d3542452c07290b1c87b98824b8ef682ec1631b44e19582f8f not found: ID does not exist" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.764777 4492 scope.go:117] "RemoveContainer" containerID="c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926" Feb 20 06:54:41 crc kubenswrapper[4492]: E0220 06:54:41.765157 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926\": container with ID starting with c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926 not found: ID does not exist" containerID="c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926" Feb 20 06:54:41 crc kubenswrapper[4492]: I0220 06:54:41.765193 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926"} err="failed to get container status \"c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926\": rpc error: code = NotFound desc = could not find container \"c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926\": container with ID starting with c01869d6018ec029ecc52fcca7dc5b04fd2e929019970eaa7fa540039cfb4926 not found: ID does not exist" Feb 20 06:54:42 crc kubenswrapper[4492]: I0220 06:54:42.293964 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 20 06:54:43 crc kubenswrapper[4492]: I0220 06:54:43.598381 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" path="/var/lib/kubelet/pods/a88f30d1-ceb3-423a-bcc9-ef66643cd6d5/volumes" Feb 20 06:54:43 crc kubenswrapper[4492]: I0220 06:54:43.599998 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" path="/var/lib/kubelet/pods/fdf66c08-5b42-418e-b48b-d2f32e6830f7/volumes" Feb 20 06:54:43 crc kubenswrapper[4492]: E0220 06:54:43.807656 4492 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 192.168.26.25:38650->192.168.26.25:44007: read tcp 192.168.26.25:38650->192.168.26.25:44007: read: connection reset by peer Feb 20 06:54:47 crc kubenswrapper[4492]: I0220 06:54:47.231914 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:47 crc kubenswrapper[4492]: I0220 06:54:47.232324 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:47 crc kubenswrapper[4492]: I0220 06:54:47.301147 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:47 crc kubenswrapper[4492]: I0220 06:54:47.817354 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.367133 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.609650 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.783955 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.835234 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b55c8b857-7s4rc"] Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.850982 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbdd3644-632d-4bf1-90d0-0fb9a3392884","Type":"ContainerStarted","Data":"d5d5584e72643889a05fd8748324f40b8378da6a6c6b3d00addcf3c66a060268"} Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.851179 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" podUID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerName="dnsmasq-dns" containerID="cri-o://1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac" gracePeriod=10 Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.912035 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6689d99595-6hcn9"] Feb 20 06:54:49 crc kubenswrapper[4492]: E0220 06:54:49.925448 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerName="init" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.925500 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerName="init" Feb 20 06:54:49 crc kubenswrapper[4492]: E0220 06:54:49.925522 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerName="dnsmasq-dns" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.925528 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerName="dnsmasq-dns" Feb 20 06:54:49 crc kubenswrapper[4492]: E0220 06:54:49.925536 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerName="dnsmasq-dns" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.925542 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerName="dnsmasq-dns" Feb 20 06:54:49 crc kubenswrapper[4492]: E0220 06:54:49.925554 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerName="init" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.925560 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerName="init" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.925858 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf66c08-5b42-418e-b48b-d2f32e6830f7" containerName="dnsmasq-dns" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.925883 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a88f30d1-ceb3-423a-bcc9-ef66643cd6d5" containerName="dnsmasq-dns" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.926871 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:49 crc kubenswrapper[4492]: I0220 06:54:49.953376 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6689d99595-6hcn9"] Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.025373 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-sb\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.025455 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-dns-svc\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.025652 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfm7x\" (UniqueName: \"kubernetes.io/projected/6f419e29-5326-4666-be72-88e790d151f5-kube-api-access-lfm7x\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.025691 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-config\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.025750 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-nb\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.127550 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-dns-svc\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.127636 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfm7x\" (UniqueName: \"kubernetes.io/projected/6f419e29-5326-4666-be72-88e790d151f5-kube-api-access-lfm7x\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.127666 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-config\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.127697 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-nb\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.127758 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-sb\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.128970 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-sb\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.129252 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-dns-svc\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.129324 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-config\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.129726 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-nb\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.144875 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfm7x\" (UniqueName: \"kubernetes.io/projected/6f419e29-5326-4666-be72-88e790d151f5-kube-api-access-lfm7x\") pod \"dnsmasq-dns-6689d99595-6hcn9\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.277601 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.299191 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.433161 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-dns-svc\") pod \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.433242 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-config\") pod \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.433348 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-226ns\" (UniqueName: \"kubernetes.io/projected/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-kube-api-access-226ns\") pod \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.433511 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb\") pod \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.606961 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-kube-api-access-226ns" (OuterVolumeSpecName: "kube-api-access-226ns") pod "11ee4e73-dcc1-42d5-9b1f-c1a55935168c" (UID: "11ee4e73-dcc1-42d5-9b1f-c1a55935168c"). InnerVolumeSpecName "kube-api-access-226ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.636593 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-226ns\" (UniqueName: \"kubernetes.io/projected/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-kube-api-access-226ns\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.841019 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11ee4e73-dcc1-42d5-9b1f-c1a55935168c" (UID: "11ee4e73-dcc1-42d5-9b1f-c1a55935168c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.841818 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb\") pod \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\" (UID: \"11ee4e73-dcc1-42d5-9b1f-c1a55935168c\") " Feb 20 06:54:50 crc kubenswrapper[4492]: W0220 06:54:50.841918 4492 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/11ee4e73-dcc1-42d5-9b1f-c1a55935168c/volumes/kubernetes.io~configmap/ovsdbserver-nb Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.841932 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11ee4e73-dcc1-42d5-9b1f-c1a55935168c" (UID: "11ee4e73-dcc1-42d5-9b1f-c1a55935168c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.842312 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.848308 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11ee4e73-dcc1-42d5-9b1f-c1a55935168c" (UID: "11ee4e73-dcc1-42d5-9b1f-c1a55935168c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.857083 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-config" (OuterVolumeSpecName: "config") pod "11ee4e73-dcc1-42d5-9b1f-c1a55935168c" (UID: "11ee4e73-dcc1-42d5-9b1f-c1a55935168c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.865729 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c936d390-aaa6-4904-a1ca-357e2dac4cde","Type":"ContainerStarted","Data":"6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.867257 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7csh" event={"ID":"3778ef17-0366-4509-8b07-64c301e9de60","Type":"ContainerStarted","Data":"b08e370736d48e767274f4045355e254bb902c3ed44fc154d82597dc6d75c44d"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.868393 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a2719d90-a43e-4b5a-b4ad-e948f75c763a","Type":"ContainerStarted","Data":"89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.870204 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a40f3503-0190-40c1-bbfb-17a454347633","Type":"ContainerStarted","Data":"adb8d227ec4946ece360c5efdde755b7684dfe30570331905fd9f555853070ac"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.872180 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k7g9" event={"ID":"53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb","Type":"ContainerStarted","Data":"2726bd74cc265fce78bddfb428ae4878cc8562adda1b20ecf4f4eedadfb744aa"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.872613 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6k7g9" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.876281 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3b8d9698-66bb-44b6-b225-e77fc088ff02","Type":"ContainerStarted","Data":"4d96848c16ecf271149d42fd6641d518d012a4699f87e234c6ad8742d172175f"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.876313 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3b8d9698-66bb-44b6-b225-e77fc088ff02","Type":"ContainerStarted","Data":"19e4a974bbf41d80a160681edadf368070bbfa52734cc92c435b7d3c2b2a7909"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.877877 4492 generic.go:334] "Generic (PLEG): container finished" podID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerID="1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac" exitCode=0 Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.877919 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" event={"ID":"11ee4e73-dcc1-42d5-9b1f-c1a55935168c","Type":"ContainerDied","Data":"1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.877944 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" event={"ID":"11ee4e73-dcc1-42d5-9b1f-c1a55935168c","Type":"ContainerDied","Data":"7e6015c5295577b98bc629b3034ebccc89b8e9f9635289add1aa27280359b3e3"} Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.877962 4492 scope.go:117] "RemoveContainer" containerID="1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.878092 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b55c8b857-7s4rc" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.901347 4492 scope.go:117] "RemoveContainer" containerID="bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.928184 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6689d99595-6hcn9"] Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.932371 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 20 06:54:50 crc kubenswrapper[4492]: E0220 06:54:50.932803 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerName="init" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.932879 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerName="init" Feb 20 06:54:50 crc kubenswrapper[4492]: E0220 06:54:50.932945 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerName="dnsmasq-dns" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.932994 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerName="dnsmasq-dns" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.933171 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" containerName="dnsmasq-dns" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.936821 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.939513 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.940536 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.940836 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.943887 4492 scope.go:117] "RemoveContainer" containerID="1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.944197 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-fbxwr" Feb 20 06:54:50 crc kubenswrapper[4492]: E0220 06:54:50.944774 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac\": container with ID starting with 1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac not found: ID does not exist" containerID="1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.944873 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac"} err="failed to get container status \"1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac\": rpc error: code = NotFound desc = could not find container \"1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac\": container with ID starting with 1b5a60d363a37a18a95fabe02dc781200bb3e3190bb1533e8ee524ad7a65e3ac not found: ID does not exist" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.944944 4492 scope.go:117] "RemoveContainer" containerID="bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.945998 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.946080 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11ee4e73-dcc1-42d5-9b1f-c1a55935168c-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:50 crc kubenswrapper[4492]: E0220 06:54:50.946154 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb\": container with ID starting with bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb not found: ID does not exist" containerID="bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.946212 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb"} err="failed to get container status \"bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb\": rpc error: code = NotFound desc = could not find container \"bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb\": container with ID starting with bc2bee5f2b59b1107c9c15e541fd27723e043444f242c689b2aec57f6bac4acb not found: ID does not exist" Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.976030 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 20 06:54:50 crc kubenswrapper[4492]: I0220 06:54:50.993629 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=24.644729367 podStartE2EDuration="38.993599838s" podCreationTimestamp="2026-02-20 06:54:12 +0000 UTC" firstStartedPulling="2026-02-20 06:54:34.99823154 +0000 UTC m=+831.769520508" lastFinishedPulling="2026-02-20 06:54:49.347102001 +0000 UTC m=+846.118390979" observedRunningTime="2026-02-20 06:54:50.981543277 +0000 UTC m=+847.752832256" watchObservedRunningTime="2026-02-20 06:54:50.993599838 +0000 UTC m=+847.764888817" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.013044 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6k7g9" podStartSLOduration=20.858370415 podStartE2EDuration="39.013025988s" podCreationTimestamp="2026-02-20 06:54:12 +0000 UTC" firstStartedPulling="2026-02-20 06:54:31.193126209 +0000 UTC m=+827.964415187" lastFinishedPulling="2026-02-20 06:54:49.347781783 +0000 UTC m=+846.119070760" observedRunningTime="2026-02-20 06:54:51.009277864 +0000 UTC m=+847.780566843" watchObservedRunningTime="2026-02-20 06:54:51.013025988 +0000 UTC m=+847.784314956" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.062520 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b55c8b857-7s4rc"] Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.066420 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b55c8b857-7s4rc"] Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.149594 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.149664 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk4w2\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-kube-api-access-qk4w2\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.149709 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-lock\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.149759 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.149797 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-cache\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.149888 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.252048 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-cache\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.252137 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.252298 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.252343 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk4w2\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-kube-api-access-qk4w2\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.252382 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-lock\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.252427 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: E0220 06:54:51.252637 4492 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 06:54:51 crc kubenswrapper[4492]: E0220 06:54:51.252660 4492 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 06:54:51 crc kubenswrapper[4492]: E0220 06:54:51.252712 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift podName:12f20cf1-c22c-48a2-9d5f-45f02009d0a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:54:51.7526917 +0000 UTC m=+848.523980678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift") pod "swift-storage-0" (UID: "12f20cf1-c22c-48a2-9d5f-45f02009d0a5") : configmap "swift-ring-files" not found Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.253231 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-cache\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.254061 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-lock\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.254146 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.259144 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.273317 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk4w2\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-kube-api-access-qk4w2\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.275301 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.458011 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-sftjr"] Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.459932 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.461439 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.463052 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.463666 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.490353 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-sftjr"] Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.558831 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-dispersionconf\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.559045 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-scripts\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.559112 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-combined-ca-bundle\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.559156 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-ring-data-devices\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.559350 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc4qc\" (UniqueName: \"kubernetes.io/projected/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-kube-api-access-rc4qc\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.559454 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-swiftconf\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.559849 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-etc-swift\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.570294 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ee4e73-dcc1-42d5-9b1f-c1a55935168c" path="/var/lib/kubelet/pods/11ee4e73-dcc1-42d5-9b1f-c1a55935168c/volumes" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.661264 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-dispersionconf\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.662288 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-scripts\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.662318 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-combined-ca-bundle\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.662349 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-ring-data-devices\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.662465 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc4qc\" (UniqueName: \"kubernetes.io/projected/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-kube-api-access-rc4qc\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.662516 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-swiftconf\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.662636 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-etc-swift\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.663076 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-etc-swift\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.663333 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-scripts\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.663729 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-ring-data-devices\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.676980 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-dispersionconf\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.682981 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-swiftconf\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.683569 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-combined-ca-bundle\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.686464 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc4qc\" (UniqueName: \"kubernetes.io/projected/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-kube-api-access-rc4qc\") pod \"swift-ring-rebalance-sftjr\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.764642 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:51 crc kubenswrapper[4492]: E0220 06:54:51.764834 4492 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 06:54:51 crc kubenswrapper[4492]: E0220 06:54:51.764865 4492 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 06:54:51 crc kubenswrapper[4492]: E0220 06:54:51.764932 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift podName:12f20cf1-c22c-48a2-9d5f-45f02009d0a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:54:52.764913838 +0000 UTC m=+849.536202816 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift") pod "swift-storage-0" (UID: "12f20cf1-c22c-48a2-9d5f-45f02009d0a5") : configmap "swift-ring-files" not found Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.776545 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.893105 4492 generic.go:334] "Generic (PLEG): container finished" podID="6f419e29-5326-4666-be72-88e790d151f5" containerID="912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d" exitCode=0 Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.893219 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" event={"ID":"6f419e29-5326-4666-be72-88e790d151f5","Type":"ContainerDied","Data":"912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d"} Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.893265 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" event={"ID":"6f419e29-5326-4666-be72-88e790d151f5","Type":"ContainerStarted","Data":"c5b01d3622f6dbb8179076fe1739f0e1eeda51eaa49b8b7227cc00612d8a6015"} Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.909969 4492 generic.go:334] "Generic (PLEG): container finished" podID="3778ef17-0366-4509-8b07-64c301e9de60" containerID="b08e370736d48e767274f4045355e254bb902c3ed44fc154d82597dc6d75c44d" exitCode=0 Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.910073 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7csh" event={"ID":"3778ef17-0366-4509-8b07-64c301e9de60","Type":"ContainerDied","Data":"b08e370736d48e767274f4045355e254bb902c3ed44fc154d82597dc6d75c44d"} Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.930836 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"dbdd3644-632d-4bf1-90d0-0fb9a3392884","Type":"ContainerStarted","Data":"ba554f5a0ed710416ca1ac9fb89228808226c35f964ea77581ca4c16683a708f"} Feb 20 06:54:51 crc kubenswrapper[4492]: I0220 06:54:51.965615 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=23.527945301 podStartE2EDuration="36.965580014s" podCreationTimestamp="2026-02-20 06:54:15 +0000 UTC" firstStartedPulling="2026-02-20 06:54:35.91064078 +0000 UTC m=+832.681929758" lastFinishedPulling="2026-02-20 06:54:49.348275473 +0000 UTC m=+846.119564471" observedRunningTime="2026-02-20 06:54:51.963136809 +0000 UTC m=+848.734425787" watchObservedRunningTime="2026-02-20 06:54:51.965580014 +0000 UTC m=+848.736868992" Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.262248 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-sftjr"] Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.663167 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.756103 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.793457 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.795244 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:52 crc kubenswrapper[4492]: E0220 06:54:52.795466 4492 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 06:54:52 crc kubenswrapper[4492]: E0220 06:54:52.795510 4492 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 06:54:52 crc kubenswrapper[4492]: E0220 06:54:52.795579 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift podName:12f20cf1-c22c-48a2-9d5f-45f02009d0a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:54:54.795556976 +0000 UTC m=+851.566845954 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift") pod "swift-storage-0" (UID: "12f20cf1-c22c-48a2-9d5f-45f02009d0a5") : configmap "swift-ring-files" not found Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.936954 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" event={"ID":"6f419e29-5326-4666-be72-88e790d151f5","Type":"ContainerStarted","Data":"7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb"} Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.937151 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.940340 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7csh" event={"ID":"3778ef17-0366-4509-8b07-64c301e9de60","Type":"ContainerStarted","Data":"abcc0cb621e55d9c07d9a562da16ffd2eb35f09ed94631630ce9233a4a5cfa56"} Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.940408 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t7csh" event={"ID":"3778ef17-0366-4509-8b07-64c301e9de60","Type":"ContainerStarted","Data":"ceb409802fba7557e94f3e8f0b8f60df8ba98f5ba3accbad1dbcc076ef63fb03"} Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.940555 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.941718 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sftjr" event={"ID":"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5","Type":"ContainerStarted","Data":"3d97da14fa3d1191f88385e18f05017ffe831fb0dce84012ed3f95385fad9c85"} Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.942240 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:52 crc kubenswrapper[4492]: I0220 06:54:52.957892 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" podStartSLOduration=3.95787316 podStartE2EDuration="3.95787316s" podCreationTimestamp="2026-02-20 06:54:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:54:52.956273344 +0000 UTC m=+849.727562323" watchObservedRunningTime="2026-02-20 06:54:52.95787316 +0000 UTC m=+849.729162138" Feb 20 06:54:53 crc kubenswrapper[4492]: I0220 06:54:53.002392 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-t7csh" podStartSLOduration=14.845995875 podStartE2EDuration="41.002375797s" podCreationTimestamp="2026-02-20 06:54:12 +0000 UTC" firstStartedPulling="2026-02-20 06:54:23.167213076 +0000 UTC m=+819.938502053" lastFinishedPulling="2026-02-20 06:54:49.323592998 +0000 UTC m=+846.094881975" observedRunningTime="2026-02-20 06:54:52.995238185 +0000 UTC m=+849.766527164" watchObservedRunningTime="2026-02-20 06:54:53.002375797 +0000 UTC m=+849.773664775" Feb 20 06:54:53 crc kubenswrapper[4492]: I0220 06:54:53.008887 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:54:53 crc kubenswrapper[4492]: I0220 06:54:53.663700 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:54 crc kubenswrapper[4492]: I0220 06:54:54.844921 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:54 crc kubenswrapper[4492]: E0220 06:54:54.845452 4492 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 06:54:54 crc kubenswrapper[4492]: E0220 06:54:54.845862 4492 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 06:54:54 crc kubenswrapper[4492]: E0220 06:54:54.845947 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift podName:12f20cf1-c22c-48a2-9d5f-45f02009d0a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:54:58.845916281 +0000 UTC m=+855.617205259 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift") pod "swift-storage-0" (UID: "12f20cf1-c22c-48a2-9d5f-45f02009d0a5") : configmap "swift-ring-files" not found Feb 20 06:54:54 crc kubenswrapper[4492]: I0220 06:54:54.960188 4492 generic.go:334] "Generic (PLEG): container finished" podID="a40f3503-0190-40c1-bbfb-17a454347633" containerID="adb8d227ec4946ece360c5efdde755b7684dfe30570331905fd9f555853070ac" exitCode=0 Feb 20 06:54:54 crc kubenswrapper[4492]: I0220 06:54:54.961001 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a40f3503-0190-40c1-bbfb-17a454347633","Type":"ContainerDied","Data":"adb8d227ec4946ece360c5efdde755b7684dfe30570331905fd9f555853070ac"} Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.700867 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.754996 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.791571 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-gnnkh"] Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.793570 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.799034 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.834990 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gnnkh"] Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.883711 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vdwc\" (UniqueName: \"kubernetes.io/projected/352e41ab-e4e2-4269-bb49-526db86873ea-kube-api-access-7vdwc\") pod \"root-account-create-update-gnnkh\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.883985 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/352e41ab-e4e2-4269-bb49-526db86873ea-operator-scripts\") pod \"root-account-create-update-gnnkh\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.986686 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vdwc\" (UniqueName: \"kubernetes.io/projected/352e41ab-e4e2-4269-bb49-526db86873ea-kube-api-access-7vdwc\") pod \"root-account-create-update-gnnkh\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.986865 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/352e41ab-e4e2-4269-bb49-526db86873ea-operator-scripts\") pod \"root-account-create-update-gnnkh\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:55 crc kubenswrapper[4492]: I0220 06:54:55.987624 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/352e41ab-e4e2-4269-bb49-526db86873ea-operator-scripts\") pod \"root-account-create-update-gnnkh\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:56 crc kubenswrapper[4492]: I0220 06:54:56.008991 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vdwc\" (UniqueName: \"kubernetes.io/projected/352e41ab-e4e2-4269-bb49-526db86873ea-kube-api-access-7vdwc\") pod \"root-account-create-update-gnnkh\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:56 crc kubenswrapper[4492]: I0220 06:54:56.160344 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:56 crc kubenswrapper[4492]: I0220 06:54:56.790153 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.037536 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.043632 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.067532 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.067779 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.067932 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-s7xvf" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.068074 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.086761 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.117968 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.118172 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2d85\" (UniqueName: \"kubernetes.io/projected/28bcc602-c9ec-4bfe-aed6-b45f16a327db-kube-api-access-n2d85\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.118302 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28bcc602-c9ec-4bfe-aed6-b45f16a327db-config\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.118526 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.118658 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.119629 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/28bcc602-c9ec-4bfe-aed6-b45f16a327db-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.120037 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28bcc602-c9ec-4bfe-aed6-b45f16a327db-scripts\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.222357 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/28bcc602-c9ec-4bfe-aed6-b45f16a327db-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.222674 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28bcc602-c9ec-4bfe-aed6-b45f16a327db-scripts\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.222725 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.222757 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2d85\" (UniqueName: \"kubernetes.io/projected/28bcc602-c9ec-4bfe-aed6-b45f16a327db-kube-api-access-n2d85\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.222782 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28bcc602-c9ec-4bfe-aed6-b45f16a327db-config\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.222835 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.222853 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.223171 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/28bcc602-c9ec-4bfe-aed6-b45f16a327db-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.223527 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/28bcc602-c9ec-4bfe-aed6-b45f16a327db-scripts\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.224097 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28bcc602-c9ec-4bfe-aed6-b45f16a327db-config\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.228206 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.232053 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.248746 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/28bcc602-c9ec-4bfe-aed6-b45f16a327db-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.257004 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2d85\" (UniqueName: \"kubernetes.io/projected/28bcc602-c9ec-4bfe-aed6-b45f16a327db-kube-api-access-n2d85\") pod \"ovn-northd-0\" (UID: \"28bcc602-c9ec-4bfe-aed6-b45f16a327db\") " pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.351547 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-gnnkh"] Feb 20 06:54:57 crc kubenswrapper[4492]: W0220 06:54:57.360967 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod352e41ab_e4e2_4269_bb49_526db86873ea.slice/crio-dfdde62eb48e07f0f15aced864617fbc13add44fbf2ba32cee5a42290584b9a7 WatchSource:0}: Error finding container dfdde62eb48e07f0f15aced864617fbc13add44fbf2ba32cee5a42290584b9a7: Status 404 returned error can't find the container with id dfdde62eb48e07f0f15aced864617fbc13add44fbf2ba32cee5a42290584b9a7 Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.396932 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 20 06:54:57 crc kubenswrapper[4492]: I0220 06:54:57.815711 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 20 06:54:57 crc kubenswrapper[4492]: W0220 06:54:57.818828 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28bcc602_c9ec_4bfe_aed6_b45f16a327db.slice/crio-e7118f77fde54f3241c561945dfe808b5bbbfee184ff4f213f69fc4900c86d34 WatchSource:0}: Error finding container e7118f77fde54f3241c561945dfe808b5bbbfee184ff4f213f69fc4900c86d34: Status 404 returned error can't find the container with id e7118f77fde54f3241c561945dfe808b5bbbfee184ff4f213f69fc4900c86d34 Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.030640 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"28bcc602-c9ec-4bfe-aed6-b45f16a327db","Type":"ContainerStarted","Data":"e7118f77fde54f3241c561945dfe808b5bbbfee184ff4f213f69fc4900c86d34"} Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.032745 4492 generic.go:334] "Generic (PLEG): container finished" podID="352e41ab-e4e2-4269-bb49-526db86873ea" containerID="36ddd304e22b8b317e53833f8714a3f92f0c978cd473986cfd86e344942523b1" exitCode=0 Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.032878 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gnnkh" event={"ID":"352e41ab-e4e2-4269-bb49-526db86873ea","Type":"ContainerDied","Data":"36ddd304e22b8b317e53833f8714a3f92f0c978cd473986cfd86e344942523b1"} Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.032943 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gnnkh" event={"ID":"352e41ab-e4e2-4269-bb49-526db86873ea","Type":"ContainerStarted","Data":"dfdde62eb48e07f0f15aced864617fbc13add44fbf2ba32cee5a42290584b9a7"} Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.035173 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a40f3503-0190-40c1-bbfb-17a454347633","Type":"ContainerStarted","Data":"b9f56ebe3d65ea17d5f27b39b6f371914df7c5c87478ff28a16f75536acd4bdc"} Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.039222 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sftjr" event={"ID":"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5","Type":"ContainerStarted","Data":"4b13e0c8e4fddea7a80e53a49fcdd32df4bceaabe2e843c7b8b594267417eec5"} Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.081360 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371982.773441 podStartE2EDuration="54.081334538s" podCreationTimestamp="2026-02-20 06:54:04 +0000 UTC" firstStartedPulling="2026-02-20 06:54:06.228627447 +0000 UTC m=+802.999916415" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:54:58.071235626 +0000 UTC m=+854.842524604" watchObservedRunningTime="2026-02-20 06:54:58.081334538 +0000 UTC m=+854.852623516" Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.114413 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-sftjr" podStartSLOduration=2.53012168 podStartE2EDuration="7.114396705s" podCreationTimestamp="2026-02-20 06:54:51 +0000 UTC" firstStartedPulling="2026-02-20 06:54:52.276126594 +0000 UTC m=+849.047415572" lastFinishedPulling="2026-02-20 06:54:56.860401618 +0000 UTC m=+853.631690597" observedRunningTime="2026-02-20 06:54:58.105347813 +0000 UTC m=+854.876636790" watchObservedRunningTime="2026-02-20 06:54:58.114396705 +0000 UTC m=+854.885685682" Feb 20 06:54:58 crc kubenswrapper[4492]: I0220 06:54:58.852763 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:54:58 crc kubenswrapper[4492]: E0220 06:54:58.853087 4492 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 20 06:54:58 crc kubenswrapper[4492]: E0220 06:54:58.853139 4492 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 20 06:54:58 crc kubenswrapper[4492]: E0220 06:54:58.853246 4492 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift podName:12f20cf1-c22c-48a2-9d5f-45f02009d0a5 nodeName:}" failed. No retries permitted until 2026-02-20 06:55:06.853209084 +0000 UTC m=+863.624498062 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift") pod "swift-storage-0" (UID: "12f20cf1-c22c-48a2-9d5f-45f02009d0a5") : configmap "swift-ring-files" not found Feb 20 06:54:59 crc kubenswrapper[4492]: I0220 06:54:59.545034 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gnnkh" Feb 20 06:54:59 crc kubenswrapper[4492]: I0220 06:54:59.676859 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/352e41ab-e4e2-4269-bb49-526db86873ea-operator-scripts\") pod \"352e41ab-e4e2-4269-bb49-526db86873ea\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " Feb 20 06:54:59 crc kubenswrapper[4492]: I0220 06:54:59.676955 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vdwc\" (UniqueName: \"kubernetes.io/projected/352e41ab-e4e2-4269-bb49-526db86873ea-kube-api-access-7vdwc\") pod \"352e41ab-e4e2-4269-bb49-526db86873ea\" (UID: \"352e41ab-e4e2-4269-bb49-526db86873ea\") " Feb 20 06:54:59 crc kubenswrapper[4492]: I0220 06:54:59.677707 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/352e41ab-e4e2-4269-bb49-526db86873ea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "352e41ab-e4e2-4269-bb49-526db86873ea" (UID: "352e41ab-e4e2-4269-bb49-526db86873ea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:54:59 crc kubenswrapper[4492]: I0220 06:54:59.677862 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/352e41ab-e4e2-4269-bb49-526db86873ea-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:54:59 crc kubenswrapper[4492]: I0220 06:54:59.681982 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352e41ab-e4e2-4269-bb49-526db86873ea-kube-api-access-7vdwc" (OuterVolumeSpecName: "kube-api-access-7vdwc") pod "352e41ab-e4e2-4269-bb49-526db86873ea" (UID: "352e41ab-e4e2-4269-bb49-526db86873ea"). InnerVolumeSpecName "kube-api-access-7vdwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:54:59 crc kubenswrapper[4492]: I0220 06:54:59.780043 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vdwc\" (UniqueName: \"kubernetes.io/projected/352e41ab-e4e2-4269-bb49-526db86873ea-kube-api-access-7vdwc\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.083584 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"28bcc602-c9ec-4bfe-aed6-b45f16a327db","Type":"ContainerStarted","Data":"cfb3d9baf4cc7eb08c49c3ca1a7b169f215d50d85f6f13aaa3f52deb7a316c16"} Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.083651 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"28bcc602-c9ec-4bfe-aed6-b45f16a327db","Type":"ContainerStarted","Data":"ab11eaf3e083ad07ee1a4a67ace4bf081785cbba84000ed53beab3a515d7ed9b"} Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.083758 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.086395 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-gnnkh" event={"ID":"352e41ab-e4e2-4269-bb49-526db86873ea","Type":"ContainerDied","Data":"dfdde62eb48e07f0f15aced864617fbc13add44fbf2ba32cee5a42290584b9a7"} Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.086457 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfdde62eb48e07f0f15aced864617fbc13add44fbf2ba32cee5a42290584b9a7" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.086573 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-gnnkh" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.116312 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.512605538 podStartE2EDuration="4.116294651s" podCreationTimestamp="2026-02-20 06:54:56 +0000 UTC" firstStartedPulling="2026-02-20 06:54:57.822228028 +0000 UTC m=+854.593516996" lastFinishedPulling="2026-02-20 06:54:59.425917131 +0000 UTC m=+856.197206109" observedRunningTime="2026-02-20 06:55:00.105167011 +0000 UTC m=+856.876455988" watchObservedRunningTime="2026-02-20 06:55:00.116294651 +0000 UTC m=+856.887583629" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.279699 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.327994 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b55bff5-dnq8f"] Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.328521 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" podUID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerName="dnsmasq-dns" containerID="cri-o://5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee" gracePeriod=10 Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.795624 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.902236 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-dns-svc\") pod \"2558c7b3-5ab5-45d8-b183-4f74014349da\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.902324 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-sb\") pod \"2558c7b3-5ab5-45d8-b183-4f74014349da\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.902375 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z22x5\" (UniqueName: \"kubernetes.io/projected/2558c7b3-5ab5-45d8-b183-4f74014349da-kube-api-access-z22x5\") pod \"2558c7b3-5ab5-45d8-b183-4f74014349da\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.902399 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-nb\") pod \"2558c7b3-5ab5-45d8-b183-4f74014349da\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.902417 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-config\") pod \"2558c7b3-5ab5-45d8-b183-4f74014349da\" (UID: \"2558c7b3-5ab5-45d8-b183-4f74014349da\") " Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.922324 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2558c7b3-5ab5-45d8-b183-4f74014349da-kube-api-access-z22x5" (OuterVolumeSpecName: "kube-api-access-z22x5") pod "2558c7b3-5ab5-45d8-b183-4f74014349da" (UID: "2558c7b3-5ab5-45d8-b183-4f74014349da"). InnerVolumeSpecName "kube-api-access-z22x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.940796 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2558c7b3-5ab5-45d8-b183-4f74014349da" (UID: "2558c7b3-5ab5-45d8-b183-4f74014349da"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.942054 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2558c7b3-5ab5-45d8-b183-4f74014349da" (UID: "2558c7b3-5ab5-45d8-b183-4f74014349da"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.943057 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-config" (OuterVolumeSpecName: "config") pod "2558c7b3-5ab5-45d8-b183-4f74014349da" (UID: "2558c7b3-5ab5-45d8-b183-4f74014349da"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:00 crc kubenswrapper[4492]: I0220 06:55:00.943276 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2558c7b3-5ab5-45d8-b183-4f74014349da" (UID: "2558c7b3-5ab5-45d8-b183-4f74014349da"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.005724 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z22x5\" (UniqueName: \"kubernetes.io/projected/2558c7b3-5ab5-45d8-b183-4f74014349da-kube-api-access-z22x5\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.005761 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.005773 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.005786 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.005798 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2558c7b3-5ab5-45d8-b183-4f74014349da-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.098181 4492 generic.go:334] "Generic (PLEG): container finished" podID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerID="5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee" exitCode=0 Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.098304 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" event={"ID":"2558c7b3-5ab5-45d8-b183-4f74014349da","Type":"ContainerDied","Data":"5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee"} Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.098400 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" event={"ID":"2558c7b3-5ab5-45d8-b183-4f74014349da","Type":"ContainerDied","Data":"ff0e92af632f294da9441252bac37e2c493522c59179ae30c825742cd8056ff9"} Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.098279 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b55bff5-dnq8f" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.098450 4492 scope.go:117] "RemoveContainer" containerID="5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.118505 4492 scope.go:117] "RemoveContainer" containerID="8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.136035 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b55bff5-dnq8f"] Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.141214 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59b55bff5-dnq8f"] Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.150041 4492 scope.go:117] "RemoveContainer" containerID="5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee" Feb 20 06:55:01 crc kubenswrapper[4492]: E0220 06:55:01.150507 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee\": container with ID starting with 5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee not found: ID does not exist" containerID="5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.150538 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee"} err="failed to get container status \"5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee\": rpc error: code = NotFound desc = could not find container \"5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee\": container with ID starting with 5bca3031f2998de9d56b4a98b2b19b3b5e2ba9c8784c7581042e9dff62cd2cee not found: ID does not exist" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.150564 4492 scope.go:117] "RemoveContainer" containerID="8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92" Feb 20 06:55:01 crc kubenswrapper[4492]: E0220 06:55:01.150829 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92\": container with ID starting with 8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92 not found: ID does not exist" containerID="8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.150854 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92"} err="failed to get container status \"8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92\": rpc error: code = NotFound desc = could not find container \"8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92\": container with ID starting with 8b6c246f9686a15db36f743a73f70afb78a02531c825aebdc37628983334cd92 not found: ID does not exist" Feb 20 06:55:01 crc kubenswrapper[4492]: I0220 06:55:01.566736 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2558c7b3-5ab5-45d8-b183-4f74014349da" path="/var/lib/kubelet/pods/2558c7b3-5ab5-45d8-b183-4f74014349da/volumes" Feb 20 06:55:04 crc kubenswrapper[4492]: I0220 06:55:04.126697 4492 generic.go:334] "Generic (PLEG): container finished" podID="b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" containerID="4b13e0c8e4fddea7a80e53a49fcdd32df4bceaabe2e843c7b8b594267417eec5" exitCode=0 Feb 20 06:55:04 crc kubenswrapper[4492]: I0220 06:55:04.126784 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sftjr" event={"ID":"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5","Type":"ContainerDied","Data":"4b13e0c8e4fddea7a80e53a49fcdd32df4bceaabe2e843c7b8b594267417eec5"} Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.424276 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.487721 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc4qc\" (UniqueName: \"kubernetes.io/projected/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-kube-api-access-rc4qc\") pod \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.487786 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-combined-ca-bundle\") pod \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.487855 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-etc-swift\") pod \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.487886 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-swiftconf\") pod \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.488731 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-scripts\") pod \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.488813 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-ring-data-devices\") pod \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.488960 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-dispersionconf\") pod \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\" (UID: \"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5\") " Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.489187 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" (UID: "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.489825 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" (UID: "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.490278 4492 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.490573 4492 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.494887 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-kube-api-access-rc4qc" (OuterVolumeSpecName: "kube-api-access-rc4qc") pod "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" (UID: "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5"). InnerVolumeSpecName "kube-api-access-rc4qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.497037 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" (UID: "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.510511 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" (UID: "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.510652 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-scripts" (OuterVolumeSpecName: "scripts") pod "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" (UID: "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.511564 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" (UID: "b39c38ec-fe4b-407f-b4d8-b5c60049b4b5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.592873 4492 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.592907 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.592920 4492 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.592936 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc4qc\" (UniqueName: \"kubernetes.io/projected/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-kube-api-access-rc4qc\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.592950 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b39c38ec-fe4b-407f-b4d8-b5c60049b4b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.751619 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.751758 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 20 06:55:05 crc kubenswrapper[4492]: I0220 06:55:05.816585 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 20 06:55:06 crc kubenswrapper[4492]: I0220 06:55:06.143463 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-sftjr" event={"ID":"b39c38ec-fe4b-407f-b4d8-b5c60049b4b5","Type":"ContainerDied","Data":"3d97da14fa3d1191f88385e18f05017ffe831fb0dce84012ed3f95385fad9c85"} Feb 20 06:55:06 crc kubenswrapper[4492]: I0220 06:55:06.143552 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d97da14fa3d1191f88385e18f05017ffe831fb0dce84012ed3f95385fad9c85" Feb 20 06:55:06 crc kubenswrapper[4492]: I0220 06:55:06.143566 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-sftjr" Feb 20 06:55:06 crc kubenswrapper[4492]: I0220 06:55:06.247408 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 20 06:55:06 crc kubenswrapper[4492]: I0220 06:55:06.918842 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:55:06 crc kubenswrapper[4492]: I0220 06:55:06.927680 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/12f20cf1-c22c-48a2-9d5f-45f02009d0a5-etc-swift\") pod \"swift-storage-0\" (UID: \"12f20cf1-c22c-48a2-9d5f-45f02009d0a5\") " pod="openstack/swift-storage-0" Feb 20 06:55:07 crc kubenswrapper[4492]: I0220 06:55:07.203307 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 20 06:55:07 crc kubenswrapper[4492]: I0220 06:55:07.728555 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.160866 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"2a59be86eb27bc568d1c26c35cc65a7841656b37e05d21194ae0c0d141133d74"} Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.291897 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g4ndh"] Feb 20 06:55:08 crc kubenswrapper[4492]: E0220 06:55:08.292422 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" containerName="swift-ring-rebalance" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.292453 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" containerName="swift-ring-rebalance" Feb 20 06:55:08 crc kubenswrapper[4492]: E0220 06:55:08.292510 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerName="init" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.292518 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerName="init" Feb 20 06:55:08 crc kubenswrapper[4492]: E0220 06:55:08.292554 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerName="dnsmasq-dns" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.292564 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerName="dnsmasq-dns" Feb 20 06:55:08 crc kubenswrapper[4492]: E0220 06:55:08.292612 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352e41ab-e4e2-4269-bb49-526db86873ea" containerName="mariadb-account-create-update" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.292620 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="352e41ab-e4e2-4269-bb49-526db86873ea" containerName="mariadb-account-create-update" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.292843 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2558c7b3-5ab5-45d8-b183-4f74014349da" containerName="dnsmasq-dns" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.292864 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="352e41ab-e4e2-4269-bb49-526db86873ea" containerName="mariadb-account-create-update" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.292879 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b39c38ec-fe4b-407f-b4d8-b5c60049b4b5" containerName="swift-ring-rebalance" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.293649 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.305522 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g4ndh"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.347053 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6513562e-a81e-4d6f-9ba9-1d7453aec92c-operator-scripts\") pod \"keystone-db-create-g4ndh\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.347342 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4tfk\" (UniqueName: \"kubernetes.io/projected/6513562e-a81e-4d6f-9ba9-1d7453aec92c-kube-api-access-j4tfk\") pod \"keystone-db-create-g4ndh\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.402486 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-0ad5-account-create-update-7lnn9"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.404109 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.408426 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.410811 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0ad5-account-create-update-7lnn9"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.448915 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4tfk\" (UniqueName: \"kubernetes.io/projected/6513562e-a81e-4d6f-9ba9-1d7453aec92c-kube-api-access-j4tfk\") pod \"keystone-db-create-g4ndh\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.448985 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e25e689-57da-4dc7-a3e6-1aabddb8acce-operator-scripts\") pod \"keystone-0ad5-account-create-update-7lnn9\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.449078 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trc6n\" (UniqueName: \"kubernetes.io/projected/3e25e689-57da-4dc7-a3e6-1aabddb8acce-kube-api-access-trc6n\") pod \"keystone-0ad5-account-create-update-7lnn9\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.449124 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6513562e-a81e-4d6f-9ba9-1d7453aec92c-operator-scripts\") pod \"keystone-db-create-g4ndh\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.449922 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6513562e-a81e-4d6f-9ba9-1d7453aec92c-operator-scripts\") pod \"keystone-db-create-g4ndh\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.468013 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4tfk\" (UniqueName: \"kubernetes.io/projected/6513562e-a81e-4d6f-9ba9-1d7453aec92c-kube-api-access-j4tfk\") pod \"keystone-db-create-g4ndh\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.529394 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-qxtqh"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.531539 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.541330 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qxtqh"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.550384 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xslvb\" (UniqueName: \"kubernetes.io/projected/0d794150-2156-4a32-8f44-4fe9a4b45a88-kube-api-access-xslvb\") pod \"placement-db-create-qxtqh\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.550656 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d794150-2156-4a32-8f44-4fe9a4b45a88-operator-scripts\") pod \"placement-db-create-qxtqh\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.550859 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e25e689-57da-4dc7-a3e6-1aabddb8acce-operator-scripts\") pod \"keystone-0ad5-account-create-update-7lnn9\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.551004 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trc6n\" (UniqueName: \"kubernetes.io/projected/3e25e689-57da-4dc7-a3e6-1aabddb8acce-kube-api-access-trc6n\") pod \"keystone-0ad5-account-create-update-7lnn9\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.551820 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e25e689-57da-4dc7-a3e6-1aabddb8acce-operator-scripts\") pod \"keystone-0ad5-account-create-update-7lnn9\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.573287 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trc6n\" (UniqueName: \"kubernetes.io/projected/3e25e689-57da-4dc7-a3e6-1aabddb8acce-kube-api-access-trc6n\") pod \"keystone-0ad5-account-create-update-7lnn9\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.616780 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.624814 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c239-account-create-update-kfsgt"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.626189 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.628230 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.633697 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c239-account-create-update-kfsgt"] Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.651854 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxfzc\" (UniqueName: \"kubernetes.io/projected/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-kube-api-access-wxfzc\") pod \"placement-c239-account-create-update-kfsgt\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.651960 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xslvb\" (UniqueName: \"kubernetes.io/projected/0d794150-2156-4a32-8f44-4fe9a4b45a88-kube-api-access-xslvb\") pod \"placement-db-create-qxtqh\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.652024 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d794150-2156-4a32-8f44-4fe9a4b45a88-operator-scripts\") pod \"placement-db-create-qxtqh\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.652055 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-operator-scripts\") pod \"placement-c239-account-create-update-kfsgt\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.654425 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d794150-2156-4a32-8f44-4fe9a4b45a88-operator-scripts\") pod \"placement-db-create-qxtqh\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.672060 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xslvb\" (UniqueName: \"kubernetes.io/projected/0d794150-2156-4a32-8f44-4fe9a4b45a88-kube-api-access-xslvb\") pod \"placement-db-create-qxtqh\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.717735 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.753149 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-operator-scripts\") pod \"placement-c239-account-create-update-kfsgt\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.753306 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxfzc\" (UniqueName: \"kubernetes.io/projected/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-kube-api-access-wxfzc\") pod \"placement-c239-account-create-update-kfsgt\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.755704 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-operator-scripts\") pod \"placement-c239-account-create-update-kfsgt\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.770385 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxfzc\" (UniqueName: \"kubernetes.io/projected/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-kube-api-access-wxfzc\") pod \"placement-c239-account-create-update-kfsgt\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:08 crc kubenswrapper[4492]: I0220 06:55:08.849618 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:09 crc kubenswrapper[4492]: I0220 06:55:09.018077 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:09 crc kubenswrapper[4492]: I0220 06:55:09.038119 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g4ndh"] Feb 20 06:55:09 crc kubenswrapper[4492]: I0220 06:55:09.138845 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0ad5-account-create-update-7lnn9"] Feb 20 06:55:09 crc kubenswrapper[4492]: W0220 06:55:09.230976 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6513562e_a81e_4d6f_9ba9_1d7453aec92c.slice/crio-c104870f90165ff7de9e7ccbf885d0857917bd31d0238075fa5c6410a8d4f8d2 WatchSource:0}: Error finding container c104870f90165ff7de9e7ccbf885d0857917bd31d0238075fa5c6410a8d4f8d2: Status 404 returned error can't find the container with id c104870f90165ff7de9e7ccbf885d0857917bd31d0238075fa5c6410a8d4f8d2 Feb 20 06:55:09 crc kubenswrapper[4492]: W0220 06:55:09.235024 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e25e689_57da_4dc7_a3e6_1aabddb8acce.slice/crio-3a10a6fa11f851e43f88454497b7a84c386aa2bd438f1ee1c4eb78d19c02bbba WatchSource:0}: Error finding container 3a10a6fa11f851e43f88454497b7a84c386aa2bd438f1ee1c4eb78d19c02bbba: Status 404 returned error can't find the container with id 3a10a6fa11f851e43f88454497b7a84c386aa2bd438f1ee1c4eb78d19c02bbba Feb 20 06:55:09 crc kubenswrapper[4492]: I0220 06:55:09.256646 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-qxtqh"] Feb 20 06:55:09 crc kubenswrapper[4492]: W0220 06:55:09.273559 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d794150_2156_4a32_8f44_4fe9a4b45a88.slice/crio-32d02e78238a255029588bcda63adc97bd9a361585c62c28860a6595efd3e958 WatchSource:0}: Error finding container 32d02e78238a255029588bcda63adc97bd9a361585c62c28860a6595efd3e958: Status 404 returned error can't find the container with id 32d02e78238a255029588bcda63adc97bd9a361585c62c28860a6595efd3e958 Feb 20 06:55:09 crc kubenswrapper[4492]: W0220 06:55:09.708615 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4be5bc11_b1d3_4456_b3a7_060de7a1bc2f.slice/crio-1841f73e98858a4f114980abf22890373051d63ce48b6eda395b28ee37dac8a2 WatchSource:0}: Error finding container 1841f73e98858a4f114980abf22890373051d63ce48b6eda395b28ee37dac8a2: Status 404 returned error can't find the container with id 1841f73e98858a4f114980abf22890373051d63ce48b6eda395b28ee37dac8a2 Feb 20 06:55:09 crc kubenswrapper[4492]: I0220 06:55:09.712312 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c239-account-create-update-kfsgt"] Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.182067 4492 generic.go:334] "Generic (PLEG): container finished" podID="0d794150-2156-4a32-8f44-4fe9a4b45a88" containerID="33184a9d99961a80661fe2f9c2a097634967d28b467cc60def086e864a6e75bf" exitCode=0 Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.182317 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qxtqh" event={"ID":"0d794150-2156-4a32-8f44-4fe9a4b45a88","Type":"ContainerDied","Data":"33184a9d99961a80661fe2f9c2a097634967d28b467cc60def086e864a6e75bf"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.182610 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qxtqh" event={"ID":"0d794150-2156-4a32-8f44-4fe9a4b45a88","Type":"ContainerStarted","Data":"32d02e78238a255029588bcda63adc97bd9a361585c62c28860a6595efd3e958"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.185051 4492 generic.go:334] "Generic (PLEG): container finished" podID="3e25e689-57da-4dc7-a3e6-1aabddb8acce" containerID="0dfc80c67ab2eeb6dac4d419d3ee27c2b874ccef518b98f64045ea6b4346bb5c" exitCode=0 Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.185083 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0ad5-account-create-update-7lnn9" event={"ID":"3e25e689-57da-4dc7-a3e6-1aabddb8acce","Type":"ContainerDied","Data":"0dfc80c67ab2eeb6dac4d419d3ee27c2b874ccef518b98f64045ea6b4346bb5c"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.185114 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0ad5-account-create-update-7lnn9" event={"ID":"3e25e689-57da-4dc7-a3e6-1aabddb8acce","Type":"ContainerStarted","Data":"3a10a6fa11f851e43f88454497b7a84c386aa2bd438f1ee1c4eb78d19c02bbba"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.190185 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"17aebadd3baff2e0399728963ef047af0544a6d70f5654cde8c54356532a0d28"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.190247 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"4b606fd3ce9a3f24f7a7152d592e072813e5bb2baf54f002ffadb8064cdb94d1"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.190260 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"0ab1145f32902572d9062eefec69b1b3416d4f18c9bd8b22092d74a05f30666d"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.190269 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"397deaa495d2d734df8273a6263e4162e83b6b2d5b71b51c60cd2a9099dcfb55"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.191569 4492 generic.go:334] "Generic (PLEG): container finished" podID="4be5bc11-b1d3-4456-b3a7-060de7a1bc2f" containerID="f29efd2d96984f6830b5997d3c665cd0a882f0d7956243176c9841b763c39f6c" exitCode=0 Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.191650 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c239-account-create-update-kfsgt" event={"ID":"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f","Type":"ContainerDied","Data":"f29efd2d96984f6830b5997d3c665cd0a882f0d7956243176c9841b763c39f6c"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.191677 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c239-account-create-update-kfsgt" event={"ID":"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f","Type":"ContainerStarted","Data":"1841f73e98858a4f114980abf22890373051d63ce48b6eda395b28ee37dac8a2"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.194180 4492 generic.go:334] "Generic (PLEG): container finished" podID="6513562e-a81e-4d6f-9ba9-1d7453aec92c" containerID="e8c20a4208f06b525332e7679b73ca42d6f250ba8c9fab55097f421ae3dddc0c" exitCode=0 Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.194219 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g4ndh" event={"ID":"6513562e-a81e-4d6f-9ba9-1d7453aec92c","Type":"ContainerDied","Data":"e8c20a4208f06b525332e7679b73ca42d6f250ba8c9fab55097f421ae3dddc0c"} Feb 20 06:55:10 crc kubenswrapper[4492]: I0220 06:55:10.194238 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g4ndh" event={"ID":"6513562e-a81e-4d6f-9ba9-1d7453aec92c","Type":"ContainerStarted","Data":"c104870f90165ff7de9e7ccbf885d0857917bd31d0238075fa5c6410a8d4f8d2"} Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.499455 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.608544 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trc6n\" (UniqueName: \"kubernetes.io/projected/3e25e689-57da-4dc7-a3e6-1aabddb8acce-kube-api-access-trc6n\") pod \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.608746 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e25e689-57da-4dc7-a3e6-1aabddb8acce-operator-scripts\") pod \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\" (UID: \"3e25e689-57da-4dc7-a3e6-1aabddb8acce\") " Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.610702 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e25e689-57da-4dc7-a3e6-1aabddb8acce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3e25e689-57da-4dc7-a3e6-1aabddb8acce" (UID: "3e25e689-57da-4dc7-a3e6-1aabddb8acce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.618905 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e25e689-57da-4dc7-a3e6-1aabddb8acce-kube-api-access-trc6n" (OuterVolumeSpecName: "kube-api-access-trc6n") pod "3e25e689-57da-4dc7-a3e6-1aabddb8acce" (UID: "3e25e689-57da-4dc7-a3e6-1aabddb8acce"). InnerVolumeSpecName "kube-api-access-trc6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.712551 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trc6n\" (UniqueName: \"kubernetes.io/projected/3e25e689-57da-4dc7-a3e6-1aabddb8acce-kube-api-access-trc6n\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.712612 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3e25e689-57da-4dc7-a3e6-1aabddb8acce-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.979176 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.992347 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:11 crc kubenswrapper[4492]: I0220 06:55:11.995589 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.133962 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6513562e-a81e-4d6f-9ba9-1d7453aec92c-operator-scripts\") pod \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.134510 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-operator-scripts\") pod \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.134818 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxfzc\" (UniqueName: \"kubernetes.io/projected/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-kube-api-access-wxfzc\") pod \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\" (UID: \"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f\") " Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.134943 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xslvb\" (UniqueName: \"kubernetes.io/projected/0d794150-2156-4a32-8f44-4fe9a4b45a88-kube-api-access-xslvb\") pod \"0d794150-2156-4a32-8f44-4fe9a4b45a88\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.134974 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6513562e-a81e-4d6f-9ba9-1d7453aec92c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6513562e-a81e-4d6f-9ba9-1d7453aec92c" (UID: "6513562e-a81e-4d6f-9ba9-1d7453aec92c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.134978 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4be5bc11-b1d3-4456-b3a7-060de7a1bc2f" (UID: "4be5bc11-b1d3-4456-b3a7-060de7a1bc2f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.135069 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d794150-2156-4a32-8f44-4fe9a4b45a88-operator-scripts\") pod \"0d794150-2156-4a32-8f44-4fe9a4b45a88\" (UID: \"0d794150-2156-4a32-8f44-4fe9a4b45a88\") " Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.135432 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4tfk\" (UniqueName: \"kubernetes.io/projected/6513562e-a81e-4d6f-9ba9-1d7453aec92c-kube-api-access-j4tfk\") pod \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\" (UID: \"6513562e-a81e-4d6f-9ba9-1d7453aec92c\") " Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.135561 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d794150-2156-4a32-8f44-4fe9a4b45a88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d794150-2156-4a32-8f44-4fe9a4b45a88" (UID: "0d794150-2156-4a32-8f44-4fe9a4b45a88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.136488 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6513562e-a81e-4d6f-9ba9-1d7453aec92c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.136649 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.136723 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d794150-2156-4a32-8f44-4fe9a4b45a88-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.138297 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d794150-2156-4a32-8f44-4fe9a4b45a88-kube-api-access-xslvb" (OuterVolumeSpecName: "kube-api-access-xslvb") pod "0d794150-2156-4a32-8f44-4fe9a4b45a88" (UID: "0d794150-2156-4a32-8f44-4fe9a4b45a88"). InnerVolumeSpecName "kube-api-access-xslvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.138328 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-kube-api-access-wxfzc" (OuterVolumeSpecName: "kube-api-access-wxfzc") pod "4be5bc11-b1d3-4456-b3a7-060de7a1bc2f" (UID: "4be5bc11-b1d3-4456-b3a7-060de7a1bc2f"). InnerVolumeSpecName "kube-api-access-wxfzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.139106 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6513562e-a81e-4d6f-9ba9-1d7453aec92c-kube-api-access-j4tfk" (OuterVolumeSpecName: "kube-api-access-j4tfk") pod "6513562e-a81e-4d6f-9ba9-1d7453aec92c" (UID: "6513562e-a81e-4d6f-9ba9-1d7453aec92c"). InnerVolumeSpecName "kube-api-access-j4tfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.213113 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-qxtqh" event={"ID":"0d794150-2156-4a32-8f44-4fe9a4b45a88","Type":"ContainerDied","Data":"32d02e78238a255029588bcda63adc97bd9a361585c62c28860a6595efd3e958"} Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.213289 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32d02e78238a255029588bcda63adc97bd9a361585c62c28860a6595efd3e958" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.213147 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-qxtqh" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.215405 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0ad5-account-create-update-7lnn9" event={"ID":"3e25e689-57da-4dc7-a3e6-1aabddb8acce","Type":"ContainerDied","Data":"3a10a6fa11f851e43f88454497b7a84c386aa2bd438f1ee1c4eb78d19c02bbba"} Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.215450 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0ad5-account-create-update-7lnn9" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.215464 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a10a6fa11f851e43f88454497b7a84c386aa2bd438f1ee1c4eb78d19c02bbba" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.218562 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"09f08a0847100e63bd763702bb8315de7a693dbf67f228179a05abfe1525ebac"} Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.218664 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"42a2d2a3e9ea581a8ea949a91d52d4170682a04b46520dee3c5a91e6d5bed399"} Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.220083 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c239-account-create-update-kfsgt" event={"ID":"4be5bc11-b1d3-4456-b3a7-060de7a1bc2f","Type":"ContainerDied","Data":"1841f73e98858a4f114980abf22890373051d63ce48b6eda395b28ee37dac8a2"} Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.220371 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1841f73e98858a4f114980abf22890373051d63ce48b6eda395b28ee37dac8a2" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.220116 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c239-account-create-update-kfsgt" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.222547 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g4ndh" event={"ID":"6513562e-a81e-4d6f-9ba9-1d7453aec92c","Type":"ContainerDied","Data":"c104870f90165ff7de9e7ccbf885d0857917bd31d0238075fa5c6410a8d4f8d2"} Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.222577 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c104870f90165ff7de9e7ccbf885d0857917bd31d0238075fa5c6410a8d4f8d2" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.222627 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g4ndh" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.238622 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxfzc\" (UniqueName: \"kubernetes.io/projected/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f-kube-api-access-wxfzc\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.238646 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xslvb\" (UniqueName: \"kubernetes.io/projected/0d794150-2156-4a32-8f44-4fe9a4b45a88-kube-api-access-xslvb\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.238678 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4tfk\" (UniqueName: \"kubernetes.io/projected/6513562e-a81e-4d6f-9ba9-1d7453aec92c-kube-api-access-j4tfk\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.460443 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kcmd6"] Feb 20 06:55:12 crc kubenswrapper[4492]: E0220 06:55:12.461023 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6513562e-a81e-4d6f-9ba9-1d7453aec92c" containerName="mariadb-database-create" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461039 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6513562e-a81e-4d6f-9ba9-1d7453aec92c" containerName="mariadb-database-create" Feb 20 06:55:12 crc kubenswrapper[4492]: E0220 06:55:12.461063 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4be5bc11-b1d3-4456-b3a7-060de7a1bc2f" containerName="mariadb-account-create-update" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461070 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4be5bc11-b1d3-4456-b3a7-060de7a1bc2f" containerName="mariadb-account-create-update" Feb 20 06:55:12 crc kubenswrapper[4492]: E0220 06:55:12.461086 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d794150-2156-4a32-8f44-4fe9a4b45a88" containerName="mariadb-database-create" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461096 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d794150-2156-4a32-8f44-4fe9a4b45a88" containerName="mariadb-database-create" Feb 20 06:55:12 crc kubenswrapper[4492]: E0220 06:55:12.461114 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e25e689-57da-4dc7-a3e6-1aabddb8acce" containerName="mariadb-account-create-update" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461121 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e25e689-57da-4dc7-a3e6-1aabddb8acce" containerName="mariadb-account-create-update" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461350 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d794150-2156-4a32-8f44-4fe9a4b45a88" containerName="mariadb-database-create" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461369 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="4be5bc11-b1d3-4456-b3a7-060de7a1bc2f" containerName="mariadb-account-create-update" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461379 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="6513562e-a81e-4d6f-9ba9-1d7453aec92c" containerName="mariadb-database-create" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.461396 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e25e689-57da-4dc7-a3e6-1aabddb8acce" containerName="mariadb-account-create-update" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.462228 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.469695 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kcmd6"] Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.653438 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6rpg\" (UniqueName: \"kubernetes.io/projected/11551b3b-e2e7-4c28-8886-b4bfe872010e-kube-api-access-p6rpg\") pod \"glance-db-create-kcmd6\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.654659 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11551b3b-e2e7-4c28-8886-b4bfe872010e-operator-scripts\") pod \"glance-db-create-kcmd6\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.665946 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6d7c-account-create-update-p6txs"] Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.667418 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.669754 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.691653 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6d7c-account-create-update-p6txs"] Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.756586 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6rpg\" (UniqueName: \"kubernetes.io/projected/11551b3b-e2e7-4c28-8886-b4bfe872010e-kube-api-access-p6rpg\") pod \"glance-db-create-kcmd6\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.757051 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11551b3b-e2e7-4c28-8886-b4bfe872010e-operator-scripts\") pod \"glance-db-create-kcmd6\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.759155 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11551b3b-e2e7-4c28-8886-b4bfe872010e-operator-scripts\") pod \"glance-db-create-kcmd6\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.774889 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6rpg\" (UniqueName: \"kubernetes.io/projected/11551b3b-e2e7-4c28-8886-b4bfe872010e-kube-api-access-p6rpg\") pod \"glance-db-create-kcmd6\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.794159 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.858668 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjzng\" (UniqueName: \"kubernetes.io/projected/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-kube-api-access-tjzng\") pod \"glance-6d7c-account-create-update-p6txs\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.858788 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-operator-scripts\") pod \"glance-6d7c-account-create-update-p6txs\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.963390 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjzng\" (UniqueName: \"kubernetes.io/projected/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-kube-api-access-tjzng\") pod \"glance-6d7c-account-create-update-p6txs\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.963501 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-operator-scripts\") pod \"glance-6d7c-account-create-update-p6txs\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.964339 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-operator-scripts\") pod \"glance-6d7c-account-create-update-p6txs\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.976812 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjzng\" (UniqueName: \"kubernetes.io/projected/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-kube-api-access-tjzng\") pod \"glance-6d7c-account-create-update-p6txs\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:12 crc kubenswrapper[4492]: I0220 06:55:12.989732 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:13 crc kubenswrapper[4492]: W0220 06:55:13.223879 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11551b3b_e2e7_4c28_8886_b4bfe872010e.slice/crio-690da65e457626aad56fde672a35c229531e099af079256d58f2006b1cffa705 WatchSource:0}: Error finding container 690da65e457626aad56fde672a35c229531e099af079256d58f2006b1cffa705: Status 404 returned error can't find the container with id 690da65e457626aad56fde672a35c229531e099af079256d58f2006b1cffa705 Feb 20 06:55:13 crc kubenswrapper[4492]: I0220 06:55:13.224768 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kcmd6"] Feb 20 06:55:13 crc kubenswrapper[4492]: I0220 06:55:13.240238 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"25e009770829fb7ba9fd55fe425622ab8667d6a10d035d51d7476ba12028d541"} Feb 20 06:55:13 crc kubenswrapper[4492]: I0220 06:55:13.240296 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"79005565080bc7b4cddc9f8800e15fadc2b9191b697b59afcc00c3ed1a9485f0"} Feb 20 06:55:13 crc kubenswrapper[4492]: I0220 06:55:13.396020 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6d7c-account-create-update-p6txs"] Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.265042 4492 generic.go:334] "Generic (PLEG): container finished" podID="11551b3b-e2e7-4c28-8886-b4bfe872010e" containerID="5e4128e68b709854f0df792e0dd075a47289f3b923a4db57f5da0f575394dfc5" exitCode=0 Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.265209 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kcmd6" event={"ID":"11551b3b-e2e7-4c28-8886-b4bfe872010e","Type":"ContainerDied","Data":"5e4128e68b709854f0df792e0dd075a47289f3b923a4db57f5da0f575394dfc5"} Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.265955 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kcmd6" event={"ID":"11551b3b-e2e7-4c28-8886-b4bfe872010e","Type":"ContainerStarted","Data":"690da65e457626aad56fde672a35c229531e099af079256d58f2006b1cffa705"} Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.269403 4492 generic.go:334] "Generic (PLEG): container finished" podID="d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc" containerID="aa7ec4a6bf75470292e87ceb15bb61eca26af293374afb3c45bdaad1cba963a6" exitCode=0 Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.269459 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6d7c-account-create-update-p6txs" event={"ID":"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc","Type":"ContainerDied","Data":"aa7ec4a6bf75470292e87ceb15bb61eca26af293374afb3c45bdaad1cba963a6"} Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.269510 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6d7c-account-create-update-p6txs" event={"ID":"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc","Type":"ContainerStarted","Data":"2a4728c9186fd0aeee3ec745947bfc91f36c280ff2211ad5c42614d598465534"} Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.278377 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"933bb25cd7e9c5c7a2e2a48b88dfc9a5ba9d86ae4e5597dca58665199d7e830d"} Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.338698 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-gnnkh"] Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.344459 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-gnnkh"] Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.408223 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-hhsgt"] Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.409242 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.411112 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.420990 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-hhsgt"] Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.594736 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e691f76-c662-4efa-a5e8-715c14abe03d-operator-scripts\") pod \"root-account-create-update-hhsgt\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.595308 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpc7x\" (UniqueName: \"kubernetes.io/projected/9e691f76-c662-4efa-a5e8-715c14abe03d-kube-api-access-lpc7x\") pod \"root-account-create-update-hhsgt\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.697350 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e691f76-c662-4efa-a5e8-715c14abe03d-operator-scripts\") pod \"root-account-create-update-hhsgt\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.697587 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpc7x\" (UniqueName: \"kubernetes.io/projected/9e691f76-c662-4efa-a5e8-715c14abe03d-kube-api-access-lpc7x\") pod \"root-account-create-update-hhsgt\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.698275 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e691f76-c662-4efa-a5e8-715c14abe03d-operator-scripts\") pod \"root-account-create-update-hhsgt\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.728852 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpc7x\" (UniqueName: \"kubernetes.io/projected/9e691f76-c662-4efa-a5e8-715c14abe03d-kube-api-access-lpc7x\") pod \"root-account-create-update-hhsgt\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:14 crc kubenswrapper[4492]: I0220 06:55:14.747284 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.200437 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-hhsgt"] Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.303256 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"3386b498526a5a9463881ab8bf1fa664d988084b55524754780d951280ab9e37"} Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.303310 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"75a8e1996855716c7c5e3e694c708dd389d7611827a016a9672c623d410bcced"} Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.303323 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"ba32245fdc84012e6366063a34797e2f207d6a946ff7bc1f7583604ff1d76d5a"} Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.303332 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"e4a46abc04d2425b6830322610192221425a93f531a2b688978cffe36a1ec637"} Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.306488 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hhsgt" event={"ID":"9e691f76-c662-4efa-a5e8-715c14abe03d","Type":"ContainerStarted","Data":"b33b56aa1623bedc96d88b75b367358a939db9f35d7c89dcbaf82c315449285d"} Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.569501 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352e41ab-e4e2-4269-bb49-526db86873ea" path="/var/lib/kubelet/pods/352e41ab-e4e2-4269-bb49-526db86873ea/volumes" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.689742 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.720356 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-operator-scripts\") pod \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.720499 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjzng\" (UniqueName: \"kubernetes.io/projected/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-kube-api-access-tjzng\") pod \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\" (UID: \"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc\") " Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.721687 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc" (UID: "d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.724045 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.729677 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-kube-api-access-tjzng" (OuterVolumeSpecName: "kube-api-access-tjzng") pod "d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc" (UID: "d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc"). InnerVolumeSpecName "kube-api-access-tjzng". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.770436 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.825168 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11551b3b-e2e7-4c28-8886-b4bfe872010e-operator-scripts\") pod \"11551b3b-e2e7-4c28-8886-b4bfe872010e\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.825213 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6rpg\" (UniqueName: \"kubernetes.io/projected/11551b3b-e2e7-4c28-8886-b4bfe872010e-kube-api-access-p6rpg\") pod \"11551b3b-e2e7-4c28-8886-b4bfe872010e\" (UID: \"11551b3b-e2e7-4c28-8886-b4bfe872010e\") " Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.825681 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjzng\" (UniqueName: \"kubernetes.io/projected/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc-kube-api-access-tjzng\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.826262 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11551b3b-e2e7-4c28-8886-b4bfe872010e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "11551b3b-e2e7-4c28-8886-b4bfe872010e" (UID: "11551b3b-e2e7-4c28-8886-b4bfe872010e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.828200 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11551b3b-e2e7-4c28-8886-b4bfe872010e-kube-api-access-p6rpg" (OuterVolumeSpecName: "kube-api-access-p6rpg") pod "11551b3b-e2e7-4c28-8886-b4bfe872010e" (UID: "11551b3b-e2e7-4c28-8886-b4bfe872010e"). InnerVolumeSpecName "kube-api-access-p6rpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.926337 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/11551b3b-e2e7-4c28-8886-b4bfe872010e-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:15 crc kubenswrapper[4492]: I0220 06:55:15.926365 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6rpg\" (UniqueName: \"kubernetes.io/projected/11551b3b-e2e7-4c28-8886-b4bfe872010e-kube-api-access-p6rpg\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.315918 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6d7c-account-create-update-p6txs" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.315906 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6d7c-account-create-update-p6txs" event={"ID":"d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc","Type":"ContainerDied","Data":"2a4728c9186fd0aeee3ec745947bfc91f36c280ff2211ad5c42614d598465534"} Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.316055 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a4728c9186fd0aeee3ec745947bfc91f36c280ff2211ad5c42614d598465534" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.322256 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"52a8c873876ea7ac58f32cf9366a81c9c508fa00e48654e0829ba161510ca3ae"} Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.322305 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"12f20cf1-c22c-48a2-9d5f-45f02009d0a5","Type":"ContainerStarted","Data":"1bbf0421a92dbdc1b7700624f7b8161458032ee652f7e604beaa55430f42cd37"} Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.324157 4492 generic.go:334] "Generic (PLEG): container finished" podID="9e691f76-c662-4efa-a5e8-715c14abe03d" containerID="d1212b8c0ed3fc190301b7d2e44b1802788c6ef622328e6c8751ffbf948f4305" exitCode=0 Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.324227 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hhsgt" event={"ID":"9e691f76-c662-4efa-a5e8-715c14abe03d","Type":"ContainerDied","Data":"d1212b8c0ed3fc190301b7d2e44b1802788c6ef622328e6c8751ffbf948f4305"} Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.325848 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kcmd6" event={"ID":"11551b3b-e2e7-4c28-8886-b4bfe872010e","Type":"ContainerDied","Data":"690da65e457626aad56fde672a35c229531e099af079256d58f2006b1cffa705"} Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.325881 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="690da65e457626aad56fde672a35c229531e099af079256d58f2006b1cffa705" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.325878 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kcmd6" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.364694 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.068958252 podStartE2EDuration="27.364673902s" podCreationTimestamp="2026-02-20 06:54:49 +0000 UTC" firstStartedPulling="2026-02-20 06:55:07.736989711 +0000 UTC m=+864.508278689" lastFinishedPulling="2026-02-20 06:55:14.032705361 +0000 UTC m=+870.803994339" observedRunningTime="2026-02-20 06:55:16.357463474 +0000 UTC m=+873.128752472" watchObservedRunningTime="2026-02-20 06:55:16.364673902 +0000 UTC m=+873.135962879" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.608085 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cd7cb8bfc-82g6n"] Feb 20 06:55:16 crc kubenswrapper[4492]: E0220 06:55:16.608495 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc" containerName="mariadb-account-create-update" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.608516 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc" containerName="mariadb-account-create-update" Feb 20 06:55:16 crc kubenswrapper[4492]: E0220 06:55:16.608527 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11551b3b-e2e7-4c28-8886-b4bfe872010e" containerName="mariadb-database-create" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.608533 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="11551b3b-e2e7-4c28-8886-b4bfe872010e" containerName="mariadb-database-create" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.608788 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="11551b3b-e2e7-4c28-8886-b4bfe872010e" containerName="mariadb-database-create" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.608821 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc" containerName="mariadb-account-create-update" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.609808 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.613275 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.617524 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd7cb8bfc-82g6n"] Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.641213 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.641424 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-svc\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.641590 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.641770 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-config\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.641911 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.642091 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8qtq\" (UniqueName: \"kubernetes.io/projected/c2ecdc8e-1e7a-4b25-9989-44978951e22f-kube-api-access-h8qtq\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.743801 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8qtq\" (UniqueName: \"kubernetes.io/projected/c2ecdc8e-1e7a-4b25-9989-44978951e22f-kube-api-access-h8qtq\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.743909 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.743938 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-svc\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.743982 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.744048 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-config\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.744111 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.744873 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.744930 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.744954 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.744930 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-svc\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.745309 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-config\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.759345 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8qtq\" (UniqueName: \"kubernetes.io/projected/c2ecdc8e-1e7a-4b25-9989-44978951e22f-kube-api-access-h8qtq\") pod \"dnsmasq-dns-6cd7cb8bfc-82g6n\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:16 crc kubenswrapper[4492]: I0220 06:55:16.936186 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.354629 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd7cb8bfc-82g6n"] Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.460904 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.600725 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.763147 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e691f76-c662-4efa-a5e8-715c14abe03d-operator-scripts\") pod \"9e691f76-c662-4efa-a5e8-715c14abe03d\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.763574 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpc7x\" (UniqueName: \"kubernetes.io/projected/9e691f76-c662-4efa-a5e8-715c14abe03d-kube-api-access-lpc7x\") pod \"9e691f76-c662-4efa-a5e8-715c14abe03d\" (UID: \"9e691f76-c662-4efa-a5e8-715c14abe03d\") " Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.763929 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e691f76-c662-4efa-a5e8-715c14abe03d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e691f76-c662-4efa-a5e8-715c14abe03d" (UID: "9e691f76-c662-4efa-a5e8-715c14abe03d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.764223 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e691f76-c662-4efa-a5e8-715c14abe03d-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.768570 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e691f76-c662-4efa-a5e8-715c14abe03d-kube-api-access-lpc7x" (OuterVolumeSpecName: "kube-api-access-lpc7x") pod "9e691f76-c662-4efa-a5e8-715c14abe03d" (UID: "9e691f76-c662-4efa-a5e8-715c14abe03d"). InnerVolumeSpecName "kube-api-access-lpc7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.866638 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpc7x\" (UniqueName: \"kubernetes.io/projected/9e691f76-c662-4efa-a5e8-715c14abe03d-kube-api-access-lpc7x\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.942684 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-scs24"] Feb 20 06:55:17 crc kubenswrapper[4492]: E0220 06:55:17.943109 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e691f76-c662-4efa-a5e8-715c14abe03d" containerName="mariadb-account-create-update" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.943130 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e691f76-c662-4efa-a5e8-715c14abe03d" containerName="mariadb-account-create-update" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.943382 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e691f76-c662-4efa-a5e8-715c14abe03d" containerName="mariadb-account-create-update" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.944097 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scs24" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.946138 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.947850 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jvgvq" Feb 20 06:55:17 crc kubenswrapper[4492]: I0220 06:55:17.949725 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-scs24"] Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.071165 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctlgg\" (UniqueName: \"kubernetes.io/projected/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-kube-api-access-ctlgg\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.071508 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-config-data\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.071615 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-combined-ca-bundle\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.071702 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-db-sync-config-data\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.173701 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctlgg\" (UniqueName: \"kubernetes.io/projected/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-kube-api-access-ctlgg\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.174080 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-config-data\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.174237 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-combined-ca-bundle\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.174320 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-db-sync-config-data\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.178653 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-config-data\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.178677 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-combined-ca-bundle\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.179986 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-db-sync-config-data\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.190739 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctlgg\" (UniqueName: \"kubernetes.io/projected/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-kube-api-access-ctlgg\") pod \"glance-db-sync-scs24\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.267984 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scs24" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.347419 4492 generic.go:334] "Generic (PLEG): container finished" podID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerID="412a0e64dc08264ca9bd0d302ca74fe41e028d5c9c7ba8cca05550330285047f" exitCode=0 Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.347531 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" event={"ID":"c2ecdc8e-1e7a-4b25-9989-44978951e22f","Type":"ContainerDied","Data":"412a0e64dc08264ca9bd0d302ca74fe41e028d5c9c7ba8cca05550330285047f"} Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.347678 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" event={"ID":"c2ecdc8e-1e7a-4b25-9989-44978951e22f","Type":"ContainerStarted","Data":"f9cb80cf7e788372fc64c0a0a481b5f4867bf0074f1fe7ba43ce12d4a8668bab"} Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.349981 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-hhsgt" event={"ID":"9e691f76-c662-4efa-a5e8-715c14abe03d","Type":"ContainerDied","Data":"b33b56aa1623bedc96d88b75b367358a939db9f35d7c89dcbaf82c315449285d"} Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.350035 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b33b56aa1623bedc96d88b75b367358a939db9f35d7c89dcbaf82c315449285d" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.350090 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-hhsgt" Feb 20 06:55:18 crc kubenswrapper[4492]: I0220 06:55:18.788398 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-scs24"] Feb 20 06:55:19 crc kubenswrapper[4492]: I0220 06:55:19.361300 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" event={"ID":"c2ecdc8e-1e7a-4b25-9989-44978951e22f","Type":"ContainerStarted","Data":"34b2244167dc065019507b9fa083547fab46bc50dc3f9117bf27d380d6bdb160"} Feb 20 06:55:19 crc kubenswrapper[4492]: I0220 06:55:19.362197 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:19 crc kubenswrapper[4492]: I0220 06:55:19.362808 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scs24" event={"ID":"8e4ff2bf-0f92-4271-bc11-e07604d93c5b","Type":"ContainerStarted","Data":"16235fca0bf2919e8ef631200bfa6d8bd998d5a64cc5e5add2719bfe9176d1a7"} Feb 20 06:55:19 crc kubenswrapper[4492]: I0220 06:55:19.378640 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" podStartSLOduration=3.378620846 podStartE2EDuration="3.378620846s" podCreationTimestamp="2026-02-20 06:55:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:55:19.376735773 +0000 UTC m=+876.148024761" watchObservedRunningTime="2026-02-20 06:55:19.378620846 +0000 UTC m=+876.149909824" Feb 20 06:55:22 crc kubenswrapper[4492]: I0220 06:55:22.392953 4492 generic.go:334] "Generic (PLEG): container finished" podID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerID="6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449" exitCode=0 Feb 20 06:55:22 crc kubenswrapper[4492]: I0220 06:55:22.393054 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c936d390-aaa6-4904-a1ca-357e2dac4cde","Type":"ContainerDied","Data":"6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449"} Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.040397 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6k7g9" podUID="53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb" containerName="ovn-controller" probeResult="failure" output=< Feb 20 06:55:23 crc kubenswrapper[4492]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 20 06:55:23 crc kubenswrapper[4492]: > Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.042321 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.046835 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t7csh" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.276980 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6k7g9-config-78m69"] Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.278142 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.287028 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.301359 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k7g9-config-78m69"] Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.391525 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwhfn\" (UniqueName: \"kubernetes.io/projected/06e1c81f-8010-48ed-af41-a328f51122f1-kube-api-access-zwhfn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.391621 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-additional-scripts\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.391781 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-log-ovn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.391913 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-scripts\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.391983 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.392081 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run-ovn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.403348 4492 generic.go:334] "Generic (PLEG): container finished" podID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerID="89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a" exitCode=0 Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.403378 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a2719d90-a43e-4b5a-b4ad-e948f75c763a","Type":"ContainerDied","Data":"89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a"} Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.406567 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c936d390-aaa6-4904-a1ca-357e2dac4cde","Type":"ContainerStarted","Data":"9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf"} Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.407399 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.472851 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.343589775 podStartE2EDuration="1m20.47283692s" podCreationTimestamp="2026-02-20 06:54:03 +0000 UTC" firstStartedPulling="2026-02-20 06:54:05.219181015 +0000 UTC m=+801.990469993" lastFinishedPulling="2026-02-20 06:54:49.348428161 +0000 UTC m=+846.119717138" observedRunningTime="2026-02-20 06:55:23.466849335 +0000 UTC m=+880.238138303" watchObservedRunningTime="2026-02-20 06:55:23.47283692 +0000 UTC m=+880.244125888" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.495923 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.496039 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run-ovn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.496160 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwhfn\" (UniqueName: \"kubernetes.io/projected/06e1c81f-8010-48ed-af41-a328f51122f1-kube-api-access-zwhfn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.496237 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-additional-scripts\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.496331 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-log-ovn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.496418 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-scripts\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.499155 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-scripts\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.499271 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.499353 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run-ovn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.499419 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-log-ovn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.499766 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-additional-scripts\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.523985 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwhfn\" (UniqueName: \"kubernetes.io/projected/06e1c81f-8010-48ed-af41-a328f51122f1-kube-api-access-zwhfn\") pod \"ovn-controller-6k7g9-config-78m69\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:23 crc kubenswrapper[4492]: I0220 06:55:23.593684 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:24 crc kubenswrapper[4492]: I0220 06:55:24.049566 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6k7g9-config-78m69"] Feb 20 06:55:24 crc kubenswrapper[4492]: W0220 06:55:24.068053 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06e1c81f_8010_48ed_af41_a328f51122f1.slice/crio-42bfbf4ccf56ab5f4d3bc34565414f1dc502f71d7fc7e059bf64c9b6124ed051 WatchSource:0}: Error finding container 42bfbf4ccf56ab5f4d3bc34565414f1dc502f71d7fc7e059bf64c9b6124ed051: Status 404 returned error can't find the container with id 42bfbf4ccf56ab5f4d3bc34565414f1dc502f71d7fc7e059bf64c9b6124ed051 Feb 20 06:55:24 crc kubenswrapper[4492]: I0220 06:55:24.418610 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k7g9-config-78m69" event={"ID":"06e1c81f-8010-48ed-af41-a328f51122f1","Type":"ContainerStarted","Data":"a7281b184eed81c267362a89b0acaa26c369b789487bef03cf8be5de82cd79fb"} Feb 20 06:55:24 crc kubenswrapper[4492]: I0220 06:55:24.418958 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k7g9-config-78m69" event={"ID":"06e1c81f-8010-48ed-af41-a328f51122f1","Type":"ContainerStarted","Data":"42bfbf4ccf56ab5f4d3bc34565414f1dc502f71d7fc7e059bf64c9b6124ed051"} Feb 20 06:55:24 crc kubenswrapper[4492]: I0220 06:55:24.422354 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a2719d90-a43e-4b5a-b4ad-e948f75c763a","Type":"ContainerStarted","Data":"3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b"} Feb 20 06:55:24 crc kubenswrapper[4492]: I0220 06:55:24.422746 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 20 06:55:24 crc kubenswrapper[4492]: I0220 06:55:24.460622 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6k7g9-config-78m69" podStartSLOduration=1.460609747 podStartE2EDuration="1.460609747s" podCreationTimestamp="2026-02-20 06:55:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:55:24.454494682 +0000 UTC m=+881.225783650" watchObservedRunningTime="2026-02-20 06:55:24.460609747 +0000 UTC m=+881.231898725" Feb 20 06:55:24 crc kubenswrapper[4492]: I0220 06:55:24.479949 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371954.374844 podStartE2EDuration="1m22.479931339s" podCreationTimestamp="2026-02-20 06:54:02 +0000 UTC" firstStartedPulling="2026-02-20 06:54:05.08306622 +0000 UTC m=+801.854355198" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:55:24.47779333 +0000 UTC m=+881.249082308" watchObservedRunningTime="2026-02-20 06:55:24.479931339 +0000 UTC m=+881.251220308" Feb 20 06:55:25 crc kubenswrapper[4492]: I0220 06:55:25.430900 4492 generic.go:334] "Generic (PLEG): container finished" podID="06e1c81f-8010-48ed-af41-a328f51122f1" containerID="a7281b184eed81c267362a89b0acaa26c369b789487bef03cf8be5de82cd79fb" exitCode=0 Feb 20 06:55:25 crc kubenswrapper[4492]: I0220 06:55:25.430955 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k7g9-config-78m69" event={"ID":"06e1c81f-8010-48ed-af41-a328f51122f1","Type":"ContainerDied","Data":"a7281b184eed81c267362a89b0acaa26c369b789487bef03cf8be5de82cd79fb"} Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.749888 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865660 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run-ovn\") pod \"06e1c81f-8010-48ed-af41-a328f51122f1\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865706 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run\") pod \"06e1c81f-8010-48ed-af41-a328f51122f1\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865735 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-scripts\") pod \"06e1c81f-8010-48ed-af41-a328f51122f1\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865783 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-additional-scripts\") pod \"06e1c81f-8010-48ed-af41-a328f51122f1\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865783 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "06e1c81f-8010-48ed-af41-a328f51122f1" (UID: "06e1c81f-8010-48ed-af41-a328f51122f1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865826 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwhfn\" (UniqueName: \"kubernetes.io/projected/06e1c81f-8010-48ed-af41-a328f51122f1-kube-api-access-zwhfn\") pod \"06e1c81f-8010-48ed-af41-a328f51122f1\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865824 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run" (OuterVolumeSpecName: "var-run") pod "06e1c81f-8010-48ed-af41-a328f51122f1" (UID: "06e1c81f-8010-48ed-af41-a328f51122f1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.865954 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-log-ovn\") pod \"06e1c81f-8010-48ed-af41-a328f51122f1\" (UID: \"06e1c81f-8010-48ed-af41-a328f51122f1\") " Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.866051 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "06e1c81f-8010-48ed-af41-a328f51122f1" (UID: "06e1c81f-8010-48ed-af41-a328f51122f1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.866587 4492 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.866611 4492 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-run\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.866620 4492 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/06e1c81f-8010-48ed-af41-a328f51122f1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.866854 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-scripts" (OuterVolumeSpecName: "scripts") pod "06e1c81f-8010-48ed-af41-a328f51122f1" (UID: "06e1c81f-8010-48ed-af41-a328f51122f1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.867249 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "06e1c81f-8010-48ed-af41-a328f51122f1" (UID: "06e1c81f-8010-48ed-af41-a328f51122f1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.871770 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e1c81f-8010-48ed-af41-a328f51122f1-kube-api-access-zwhfn" (OuterVolumeSpecName: "kube-api-access-zwhfn") pod "06e1c81f-8010-48ed-af41-a328f51122f1" (UID: "06e1c81f-8010-48ed-af41-a328f51122f1"). InnerVolumeSpecName "kube-api-access-zwhfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.938701 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.968891 4492 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.968924 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwhfn\" (UniqueName: \"kubernetes.io/projected/06e1c81f-8010-48ed-af41-a328f51122f1-kube-api-access-zwhfn\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:26 crc kubenswrapper[4492]: I0220 06:55:26.968938 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/06e1c81f-8010-48ed-af41-a328f51122f1-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.001059 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6689d99595-6hcn9"] Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.001412 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" podUID="6f419e29-5326-4666-be72-88e790d151f5" containerName="dnsmasq-dns" containerID="cri-o://7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb" gracePeriod=10 Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.392203 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.465465 4492 generic.go:334] "Generic (PLEG): container finished" podID="6f419e29-5326-4666-be72-88e790d151f5" containerID="7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb" exitCode=0 Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.465703 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" event={"ID":"6f419e29-5326-4666-be72-88e790d151f5","Type":"ContainerDied","Data":"7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb"} Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.465737 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" event={"ID":"6f419e29-5326-4666-be72-88e790d151f5","Type":"ContainerDied","Data":"c5b01d3622f6dbb8179076fe1739f0e1eeda51eaa49b8b7227cc00612d8a6015"} Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.465759 4492 scope.go:117] "RemoveContainer" containerID="7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.465878 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6689d99595-6hcn9" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.475365 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6k7g9-config-78m69" event={"ID":"06e1c81f-8010-48ed-af41-a328f51122f1","Type":"ContainerDied","Data":"42bfbf4ccf56ab5f4d3bc34565414f1dc502f71d7fc7e059bf64c9b6124ed051"} Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.475434 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42bfbf4ccf56ab5f4d3bc34565414f1dc502f71d7fc7e059bf64c9b6124ed051" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.475572 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6k7g9-config-78m69" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.479622 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-dns-svc\") pod \"6f419e29-5326-4666-be72-88e790d151f5\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.479702 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfm7x\" (UniqueName: \"kubernetes.io/projected/6f419e29-5326-4666-be72-88e790d151f5-kube-api-access-lfm7x\") pod \"6f419e29-5326-4666-be72-88e790d151f5\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.479793 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-config\") pod \"6f419e29-5326-4666-be72-88e790d151f5\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.479871 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-sb\") pod \"6f419e29-5326-4666-be72-88e790d151f5\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.479915 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-nb\") pod \"6f419e29-5326-4666-be72-88e790d151f5\" (UID: \"6f419e29-5326-4666-be72-88e790d151f5\") " Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.487644 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f419e29-5326-4666-be72-88e790d151f5-kube-api-access-lfm7x" (OuterVolumeSpecName: "kube-api-access-lfm7x") pod "6f419e29-5326-4666-be72-88e790d151f5" (UID: "6f419e29-5326-4666-be72-88e790d151f5"). InnerVolumeSpecName "kube-api-access-lfm7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.547845 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f419e29-5326-4666-be72-88e790d151f5" (UID: "6f419e29-5326-4666-be72-88e790d151f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.548329 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6f419e29-5326-4666-be72-88e790d151f5" (UID: "6f419e29-5326-4666-be72-88e790d151f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.555190 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6f419e29-5326-4666-be72-88e790d151f5" (UID: "6f419e29-5326-4666-be72-88e790d151f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.564721 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-config" (OuterVolumeSpecName: "config") pod "6f419e29-5326-4666-be72-88e790d151f5" (UID: "6f419e29-5326-4666-be72-88e790d151f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.570279 4492 scope.go:117] "RemoveContainer" containerID="912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.587430 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.587501 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfm7x\" (UniqueName: \"kubernetes.io/projected/6f419e29-5326-4666-be72-88e790d151f5-kube-api-access-lfm7x\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.587523 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.587538 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.587550 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f419e29-5326-4666-be72-88e790d151f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.612298 4492 scope.go:117] "RemoveContainer" containerID="7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb" Feb 20 06:55:27 crc kubenswrapper[4492]: E0220 06:55:27.612764 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb\": container with ID starting with 7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb not found: ID does not exist" containerID="7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.612809 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb"} err="failed to get container status \"7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb\": rpc error: code = NotFound desc = could not find container \"7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb\": container with ID starting with 7b2ae61268d7b66c1f0da62063e804f89b25b3a103b483ec66ef3c7f646bdccb not found: ID does not exist" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.612844 4492 scope.go:117] "RemoveContainer" containerID="912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d" Feb 20 06:55:27 crc kubenswrapper[4492]: E0220 06:55:27.613512 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d\": container with ID starting with 912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d not found: ID does not exist" containerID="912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.613557 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d"} err="failed to get container status \"912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d\": rpc error: code = NotFound desc = could not find container \"912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d\": container with ID starting with 912542fde919b7361eb0882c0bdddd0ca2dbeba30b5a28098ee206df0ddf772d not found: ID does not exist" Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.793510 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6689d99595-6hcn9"] Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.799936 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6689d99595-6hcn9"] Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.873899 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6k7g9-config-78m69"] Feb 20 06:55:27 crc kubenswrapper[4492]: I0220 06:55:27.880459 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6k7g9-config-78m69"] Feb 20 06:55:28 crc kubenswrapper[4492]: I0220 06:55:28.032742 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6k7g9" Feb 20 06:55:29 crc kubenswrapper[4492]: I0220 06:55:29.566682 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06e1c81f-8010-48ed-af41-a328f51122f1" path="/var/lib/kubelet/pods/06e1c81f-8010-48ed-af41-a328f51122f1/volumes" Feb 20 06:55:29 crc kubenswrapper[4492]: I0220 06:55:29.567563 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f419e29-5326-4666-be72-88e790d151f5" path="/var/lib/kubelet/pods/6f419e29-5326-4666-be72-88e790d151f5/volumes" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.505400 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.605460 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.825904 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-5znh9"] Feb 20 06:55:34 crc kubenswrapper[4492]: E0220 06:55:34.826328 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f419e29-5326-4666-be72-88e790d151f5" containerName="dnsmasq-dns" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.826351 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f419e29-5326-4666-be72-88e790d151f5" containerName="dnsmasq-dns" Feb 20 06:55:34 crc kubenswrapper[4492]: E0220 06:55:34.826379 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e1c81f-8010-48ed-af41-a328f51122f1" containerName="ovn-config" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.826388 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e1c81f-8010-48ed-af41-a328f51122f1" containerName="ovn-config" Feb 20 06:55:34 crc kubenswrapper[4492]: E0220 06:55:34.826400 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f419e29-5326-4666-be72-88e790d151f5" containerName="init" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.826408 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f419e29-5326-4666-be72-88e790d151f5" containerName="init" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.826617 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f419e29-5326-4666-be72-88e790d151f5" containerName="dnsmasq-dns" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.826631 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e1c81f-8010-48ed-af41-a328f51122f1" containerName="ovn-config" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.827172 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.857893 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5znh9"] Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.922852 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5lb5\" (UniqueName: \"kubernetes.io/projected/21c3e515-5917-4360-8ab2-5fc920f8f537-kube-api-access-x5lb5\") pod \"cinder-db-create-5znh9\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.922941 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21c3e515-5917-4360-8ab2-5fc920f8f537-operator-scripts\") pod \"cinder-db-create-5znh9\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.967501 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-c12a-account-create-update-qx47p"] Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.968532 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.975229 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 20 06:55:34 crc kubenswrapper[4492]: I0220 06:55:34.989880 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c12a-account-create-update-qx47p"] Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.025149 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zftht\" (UniqueName: \"kubernetes.io/projected/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-kube-api-access-zftht\") pod \"cinder-c12a-account-create-update-qx47p\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.025358 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-operator-scripts\") pod \"cinder-c12a-account-create-update-qx47p\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.025442 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5lb5\" (UniqueName: \"kubernetes.io/projected/21c3e515-5917-4360-8ab2-5fc920f8f537-kube-api-access-x5lb5\") pod \"cinder-db-create-5znh9\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.025524 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21c3e515-5917-4360-8ab2-5fc920f8f537-operator-scripts\") pod \"cinder-db-create-5znh9\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.026215 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21c3e515-5917-4360-8ab2-5fc920f8f537-operator-scripts\") pod \"cinder-db-create-5znh9\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.041316 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-lq7fc"] Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.048099 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.063107 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-lq7fc"] Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.073284 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5lb5\" (UniqueName: \"kubernetes.io/projected/21c3e515-5917-4360-8ab2-5fc920f8f537-kube-api-access-x5lb5\") pod \"cinder-db-create-5znh9\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.127112 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zftht\" (UniqueName: \"kubernetes.io/projected/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-kube-api-access-zftht\") pod \"cinder-c12a-account-create-update-qx47p\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.127679 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aba8976-47dd-4f37-9e4d-e561b7f05495-operator-scripts\") pod \"heat-db-create-lq7fc\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.127827 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp6pv\" (UniqueName: \"kubernetes.io/projected/0aba8976-47dd-4f37-9e4d-e561b7f05495-kube-api-access-hp6pv\") pod \"heat-db-create-lq7fc\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.127984 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-operator-scripts\") pod \"cinder-c12a-account-create-update-qx47p\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.128631 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-operator-scripts\") pod \"cinder-c12a-account-create-update-qx47p\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.145748 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.146407 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zftht\" (UniqueName: \"kubernetes.io/projected/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-kube-api-access-zftht\") pod \"cinder-c12a-account-create-update-qx47p\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.228035 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-qjqbf"] Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.229243 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.229668 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aba8976-47dd-4f37-9e4d-e561b7f05495-operator-scripts\") pod \"heat-db-create-lq7fc\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.229722 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp6pv\" (UniqueName: \"kubernetes.io/projected/0aba8976-47dd-4f37-9e4d-e561b7f05495-kube-api-access-hp6pv\") pod \"heat-db-create-lq7fc\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.230719 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aba8976-47dd-4f37-9e4d-e561b7f05495-operator-scripts\") pod \"heat-db-create-lq7fc\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.245734 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qjqbf"] Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.267699 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-02b5-account-create-update-nlqzc"] Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.269035 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.288506 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.302363 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp6pv\" (UniqueName: \"kubernetes.io/projected/0aba8976-47dd-4f37-9e4d-e561b7f05495-kube-api-access-hp6pv\") pod \"heat-db-create-lq7fc\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.308257 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.331854 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-operator-scripts\") pod \"heat-02b5-account-create-update-nlqzc\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.331949 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvb9t\" (UniqueName: \"kubernetes.io/projected/d59d1326-3f4e-4429-93ee-4b983de7ff82-kube-api-access-zvb9t\") pod \"barbican-db-create-qjqbf\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.331995 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q859\" (UniqueName: \"kubernetes.io/projected/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-kube-api-access-2q859\") pod \"heat-02b5-account-create-update-nlqzc\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.332051 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d59d1326-3f4e-4429-93ee-4b983de7ff82-operator-scripts\") pod \"barbican-db-create-qjqbf\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.350165 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-q26vt"] Feb 20 06:55:35 crc kubenswrapper[4492]: I0220 06:55:35.351909 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.407054 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.410135 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-02b5-account-create-update-nlqzc"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.434813 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-operator-scripts\") pod \"heat-02b5-account-create-update-nlqzc\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.434884 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvb9t\" (UniqueName: \"kubernetes.io/projected/d59d1326-3f4e-4429-93ee-4b983de7ff82-kube-api-access-zvb9t\") pod \"barbican-db-create-qjqbf\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.434922 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q859\" (UniqueName: \"kubernetes.io/projected/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-kube-api-access-2q859\") pod \"heat-02b5-account-create-update-nlqzc\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.434950 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64054bf5-de18-49b3-93c5-e5081ef122f8-operator-scripts\") pod \"neutron-db-create-q26vt\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.434995 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d59d1326-3f4e-4429-93ee-4b983de7ff82-operator-scripts\") pod \"barbican-db-create-qjqbf\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.435012 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcvrv\" (UniqueName: \"kubernetes.io/projected/64054bf5-de18-49b3-93c5-e5081ef122f8-kube-api-access-tcvrv\") pod \"neutron-db-create-q26vt\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.435632 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-operator-scripts\") pod \"heat-02b5-account-create-update-nlqzc\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.436415 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d59d1326-3f4e-4429-93ee-4b983de7ff82-operator-scripts\") pod \"barbican-db-create-qjqbf\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.451637 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-q26vt"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.456868 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvb9t\" (UniqueName: \"kubernetes.io/projected/d59d1326-3f4e-4429-93ee-4b983de7ff82-kube-api-access-zvb9t\") pod \"barbican-db-create-qjqbf\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.462111 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q859\" (UniqueName: \"kubernetes.io/projected/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-kube-api-access-2q859\") pod \"heat-02b5-account-create-update-nlqzc\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.471515 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-91fd-account-create-update-68gbg"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.472924 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.475171 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.482410 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-91fd-account-create-update-68gbg"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.500917 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-x5r2g"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.502402 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.506171 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.506289 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.506405 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.510570 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4v2r8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.518528 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-x5r2g"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.523853 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7e35-account-create-update-22fl8"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.525170 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.527685 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.536926 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-combined-ca-bundle\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.536963 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7qxc\" (UniqueName: \"kubernetes.io/projected/84053541-a741-4dd3-8b4c-420cfa5da896-kube-api-access-h7qxc\") pod \"barbican-91fd-account-create-update-68gbg\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.537003 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64054bf5-de18-49b3-93c5-e5081ef122f8-operator-scripts\") pod \"neutron-db-create-q26vt\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.537049 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcvrv\" (UniqueName: \"kubernetes.io/projected/64054bf5-de18-49b3-93c5-e5081ef122f8-kube-api-access-tcvrv\") pod \"neutron-db-create-q26vt\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.537088 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-config-data\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.537143 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84053541-a741-4dd3-8b4c-420cfa5da896-operator-scripts\") pod \"barbican-91fd-account-create-update-68gbg\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.537164 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6sjs\" (UniqueName: \"kubernetes.io/projected/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-kube-api-access-m6sjs\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.543278 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64054bf5-de18-49b3-93c5-e5081ef122f8-operator-scripts\") pod \"neutron-db-create-q26vt\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.546660 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7e35-account-create-update-22fl8"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.564174 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.590537 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcvrv\" (UniqueName: \"kubernetes.io/projected/64054bf5-de18-49b3-93c5-e5081ef122f8-kube-api-access-tcvrv\") pod \"neutron-db-create-q26vt\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.620062 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.639948 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84053541-a741-4dd3-8b4c-420cfa5da896-operator-scripts\") pod \"barbican-91fd-account-create-update-68gbg\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.640059 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6sjs\" (UniqueName: \"kubernetes.io/projected/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-kube-api-access-m6sjs\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.640095 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7dnw\" (UniqueName: \"kubernetes.io/projected/2d8c5715-e9fb-4eef-994f-04aeb6b18864-kube-api-access-h7dnw\") pod \"neutron-7e35-account-create-update-22fl8\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.640841 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84053541-a741-4dd3-8b4c-420cfa5da896-operator-scripts\") pod \"barbican-91fd-account-create-update-68gbg\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.641775 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-combined-ca-bundle\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.641856 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7qxc\" (UniqueName: \"kubernetes.io/projected/84053541-a741-4dd3-8b4c-420cfa5da896-kube-api-access-h7qxc\") pod \"barbican-91fd-account-create-update-68gbg\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.642083 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8c5715-e9fb-4eef-994f-04aeb6b18864-operator-scripts\") pod \"neutron-7e35-account-create-update-22fl8\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.642137 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-config-data\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.651977 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-config-data\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.657564 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-combined-ca-bundle\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.667112 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7qxc\" (UniqueName: \"kubernetes.io/projected/84053541-a741-4dd3-8b4c-420cfa5da896-kube-api-access-h7qxc\") pod \"barbican-91fd-account-create-update-68gbg\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.667191 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6sjs\" (UniqueName: \"kubernetes.io/projected/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-kube-api-access-m6sjs\") pod \"keystone-db-sync-x5r2g\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.722662 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.743775 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8c5715-e9fb-4eef-994f-04aeb6b18864-operator-scripts\") pod \"neutron-7e35-account-create-update-22fl8\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.743855 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7dnw\" (UniqueName: \"kubernetes.io/projected/2d8c5715-e9fb-4eef-994f-04aeb6b18864-kube-api-access-h7dnw\") pod \"neutron-7e35-account-create-update-22fl8\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.744817 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8c5715-e9fb-4eef-994f-04aeb6b18864-operator-scripts\") pod \"neutron-7e35-account-create-update-22fl8\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.771893 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7dnw\" (UniqueName: \"kubernetes.io/projected/2d8c5715-e9fb-4eef-994f-04aeb6b18864-kube-api-access-h7dnw\") pod \"neutron-7e35-account-create-update-22fl8\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.793227 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.839860 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:35.868778 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.521239 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-02b5-account-create-update-nlqzc"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.563681 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-qjqbf"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.578852 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-lq7fc"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.585164 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-91fd-account-create-update-68gbg"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.593229 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-5znh9"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.605888 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-c12a-account-create-update-qx47p"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.619886 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-02b5-account-create-update-nlqzc" event={"ID":"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f","Type":"ContainerStarted","Data":"a01c92528cde8ab08dd2b43c85eb0143294012db957111e583346ce8950eb020"} Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.620884 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-q26vt"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.640355 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7e35-account-create-update-22fl8"] Feb 20 06:55:36 crc kubenswrapper[4492]: I0220 06:55:36.660252 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-x5r2g"] Feb 20 06:55:36 crc kubenswrapper[4492]: E0220 06:55:36.716506 4492 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.25:47626->192.168.26.25:44007: write tcp 192.168.26.25:47626->192.168.26.25:44007: write: connection reset by peer Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.640266 4492 generic.go:334] "Generic (PLEG): container finished" podID="64054bf5-de18-49b3-93c5-e5081ef122f8" containerID="5b1471e28cb25ff9e5cbe97b972ca8a3e60566099b74dbd28911c8e42fa1d0f0" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.640448 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q26vt" event={"ID":"64054bf5-de18-49b3-93c5-e5081ef122f8","Type":"ContainerDied","Data":"5b1471e28cb25ff9e5cbe97b972ca8a3e60566099b74dbd28911c8e42fa1d0f0"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.640774 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q26vt" event={"ID":"64054bf5-de18-49b3-93c5-e5081ef122f8","Type":"ContainerStarted","Data":"cfcdc64de16b2037e2057d5899cea30c519a4dcfdabbe32f25442b36ffc4b038"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.674293 4492 generic.go:334] "Generic (PLEG): container finished" podID="2d8c5715-e9fb-4eef-994f-04aeb6b18864" containerID="84c58b2555461bb9a56fd038c6bbf8ac21ea32306af96e0609692552c01e09f0" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.674391 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7e35-account-create-update-22fl8" event={"ID":"2d8c5715-e9fb-4eef-994f-04aeb6b18864","Type":"ContainerDied","Data":"84c58b2555461bb9a56fd038c6bbf8ac21ea32306af96e0609692552c01e09f0"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.674842 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7e35-account-create-update-22fl8" event={"ID":"2d8c5715-e9fb-4eef-994f-04aeb6b18864","Type":"ContainerStarted","Data":"c7a35f635e5a7ab346de5c21ec1121bac3936b2544cec3196c0c6d7adec2ab25"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.679044 4492 generic.go:334] "Generic (PLEG): container finished" podID="5c067dc9-f4a3-46d3-bb5f-bc8f36312a79" containerID="21fa4b0532cfd3e0efd8c7c742e61898fed119901fe4c38113a0db67b6d6c1be" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.679138 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c12a-account-create-update-qx47p" event={"ID":"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79","Type":"ContainerDied","Data":"21fa4b0532cfd3e0efd8c7c742e61898fed119901fe4c38113a0db67b6d6c1be"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.679182 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c12a-account-create-update-qx47p" event={"ID":"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79","Type":"ContainerStarted","Data":"7436484a3fe139c2d26a0eb299621147fbb1af4df6426a8d1dec41730d838d7b"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.682283 4492 generic.go:334] "Generic (PLEG): container finished" podID="b1e8d47d-5fa9-4cae-8164-8b5bf089c82f" containerID="8b98e9676fcb838bc3cccc1cd17480c8b57e9f708a1b95d7ca19ee900fe8155a" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.682344 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-02b5-account-create-update-nlqzc" event={"ID":"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f","Type":"ContainerDied","Data":"8b98e9676fcb838bc3cccc1cd17480c8b57e9f708a1b95d7ca19ee900fe8155a"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.686821 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x5r2g" event={"ID":"eaa400f5-22b0-44f7-9655-f8b3bd1f8467","Type":"ContainerStarted","Data":"351b1eddb2d6627da07875ebda350c47cd6cbb6f31d9332ad1628c681738b177"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.693380 4492 generic.go:334] "Generic (PLEG): container finished" podID="0aba8976-47dd-4f37-9e4d-e561b7f05495" containerID="2a997b0e91405df93db58d3c407de0407413b6972da3636b4babc2909f92d11f" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.693558 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lq7fc" event={"ID":"0aba8976-47dd-4f37-9e4d-e561b7f05495","Type":"ContainerDied","Data":"2a997b0e91405df93db58d3c407de0407413b6972da3636b4babc2909f92d11f"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.693609 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lq7fc" event={"ID":"0aba8976-47dd-4f37-9e4d-e561b7f05495","Type":"ContainerStarted","Data":"dd1de0a2d4feb29acfc10f005d75f39b4f301d3457fc8535eaed17e402cb972b"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.702932 4492 generic.go:334] "Generic (PLEG): container finished" podID="d59d1326-3f4e-4429-93ee-4b983de7ff82" containerID="88ff3b8feb37a1ce94ecb080e21bdfa3c389947bbd6a378c8d7617bffa8da2ed" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.703012 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qjqbf" event={"ID":"d59d1326-3f4e-4429-93ee-4b983de7ff82","Type":"ContainerDied","Data":"88ff3b8feb37a1ce94ecb080e21bdfa3c389947bbd6a378c8d7617bffa8da2ed"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.703048 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qjqbf" event={"ID":"d59d1326-3f4e-4429-93ee-4b983de7ff82","Type":"ContainerStarted","Data":"c16f52ddcd21c58660d24a0043acbf8206dab239308436af46b89327755f5ec5"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.709399 4492 generic.go:334] "Generic (PLEG): container finished" podID="84053541-a741-4dd3-8b4c-420cfa5da896" containerID="60fee276a052c1f522f17721904a700120bf9ac76abc5f17f1a4ca7c3d821c35" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.709510 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-91fd-account-create-update-68gbg" event={"ID":"84053541-a741-4dd3-8b4c-420cfa5da896","Type":"ContainerDied","Data":"60fee276a052c1f522f17721904a700120bf9ac76abc5f17f1a4ca7c3d821c35"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.709552 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-91fd-account-create-update-68gbg" event={"ID":"84053541-a741-4dd3-8b4c-420cfa5da896","Type":"ContainerStarted","Data":"f36df9abecf22d72e391e6a9cd10cb832a5ec2d3c8a6497a8a5af5a244022b22"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.717539 4492 generic.go:334] "Generic (PLEG): container finished" podID="21c3e515-5917-4360-8ab2-5fc920f8f537" containerID="db062a294b75d86880b4bdaba992eacadeecd70bf9478ecd6a4929b0c7a63593" exitCode=0 Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.717590 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5znh9" event={"ID":"21c3e515-5917-4360-8ab2-5fc920f8f537","Type":"ContainerDied","Data":"db062a294b75d86880b4bdaba992eacadeecd70bf9478ecd6a4929b0c7a63593"} Feb 20 06:55:37 crc kubenswrapper[4492]: I0220 06:55:37.717623 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5znh9" event={"ID":"21c3e515-5917-4360-8ab2-5fc920f8f537","Type":"ContainerStarted","Data":"396af04a6f81eeac1bd82d0c4708d70085d4ac6b59496194e14ca87ab01d6ab3"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.097783 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.214132 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64054bf5-de18-49b3-93c5-e5081ef122f8-operator-scripts\") pod \"64054bf5-de18-49b3-93c5-e5081ef122f8\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.214274 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcvrv\" (UniqueName: \"kubernetes.io/projected/64054bf5-de18-49b3-93c5-e5081ef122f8-kube-api-access-tcvrv\") pod \"64054bf5-de18-49b3-93c5-e5081ef122f8\" (UID: \"64054bf5-de18-49b3-93c5-e5081ef122f8\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.215996 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64054bf5-de18-49b3-93c5-e5081ef122f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64054bf5-de18-49b3-93c5-e5081ef122f8" (UID: "64054bf5-de18-49b3-93c5-e5081ef122f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.231320 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64054bf5-de18-49b3-93c5-e5081ef122f8-kube-api-access-tcvrv" (OuterVolumeSpecName: "kube-api-access-tcvrv") pod "64054bf5-de18-49b3-93c5-e5081ef122f8" (UID: "64054bf5-de18-49b3-93c5-e5081ef122f8"). InnerVolumeSpecName "kube-api-access-tcvrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.281228 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.285057 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.290237 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.311325 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.311396 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.312120 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.316702 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64054bf5-de18-49b3-93c5-e5081ef122f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.316724 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcvrv\" (UniqueName: \"kubernetes.io/projected/64054bf5-de18-49b3-93c5-e5081ef122f8-kube-api-access-tcvrv\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.332907 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.333392 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.341551 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.417912 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21c3e515-5917-4360-8ab2-5fc920f8f537-operator-scripts\") pod \"21c3e515-5917-4360-8ab2-5fc920f8f537\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.417971 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zftht\" (UniqueName: \"kubernetes.io/projected/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-kube-api-access-zftht\") pod \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.418000 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-operator-scripts\") pod \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.418017 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2q859\" (UniqueName: \"kubernetes.io/projected/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-kube-api-access-2q859\") pod \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\" (UID: \"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.418122 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aba8976-47dd-4f37-9e4d-e561b7f05495-operator-scripts\") pod \"0aba8976-47dd-4f37-9e4d-e561b7f05495\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.418252 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-operator-scripts\") pod \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\" (UID: \"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.418332 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp6pv\" (UniqueName: \"kubernetes.io/projected/0aba8976-47dd-4f37-9e4d-e561b7f05495-kube-api-access-hp6pv\") pod \"0aba8976-47dd-4f37-9e4d-e561b7f05495\" (UID: \"0aba8976-47dd-4f37-9e4d-e561b7f05495\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.418444 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5lb5\" (UniqueName: \"kubernetes.io/projected/21c3e515-5917-4360-8ab2-5fc920f8f537-kube-api-access-x5lb5\") pod \"21c3e515-5917-4360-8ab2-5fc920f8f537\" (UID: \"21c3e515-5917-4360-8ab2-5fc920f8f537\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.419316 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c067dc9-f4a3-46d3-bb5f-bc8f36312a79" (UID: "5c067dc9-f4a3-46d3-bb5f-bc8f36312a79"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.419622 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0aba8976-47dd-4f37-9e4d-e561b7f05495-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0aba8976-47dd-4f37-9e4d-e561b7f05495" (UID: "0aba8976-47dd-4f37-9e4d-e561b7f05495"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.420081 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1e8d47d-5fa9-4cae-8164-8b5bf089c82f" (UID: "b1e8d47d-5fa9-4cae-8164-8b5bf089c82f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.421871 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21c3e515-5917-4360-8ab2-5fc920f8f537-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "21c3e515-5917-4360-8ab2-5fc920f8f537" (UID: "21c3e515-5917-4360-8ab2-5fc920f8f537"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.424040 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/21c3e515-5917-4360-8ab2-5fc920f8f537-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.424067 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.424080 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aba8976-47dd-4f37-9e4d-e561b7f05495-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.424090 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.424196 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-kube-api-access-2q859" (OuterVolumeSpecName: "kube-api-access-2q859") pod "b1e8d47d-5fa9-4cae-8164-8b5bf089c82f" (UID: "b1e8d47d-5fa9-4cae-8164-8b5bf089c82f"). InnerVolumeSpecName "kube-api-access-2q859". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.436586 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-kube-api-access-zftht" (OuterVolumeSpecName: "kube-api-access-zftht") pod "5c067dc9-f4a3-46d3-bb5f-bc8f36312a79" (UID: "5c067dc9-f4a3-46d3-bb5f-bc8f36312a79"). InnerVolumeSpecName "kube-api-access-zftht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.440622 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21c3e515-5917-4360-8ab2-5fc920f8f537-kube-api-access-x5lb5" (OuterVolumeSpecName: "kube-api-access-x5lb5") pod "21c3e515-5917-4360-8ab2-5fc920f8f537" (UID: "21c3e515-5917-4360-8ab2-5fc920f8f537"). InnerVolumeSpecName "kube-api-access-x5lb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.448867 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aba8976-47dd-4f37-9e4d-e561b7f05495-kube-api-access-hp6pv" (OuterVolumeSpecName: "kube-api-access-hp6pv") pod "0aba8976-47dd-4f37-9e4d-e561b7f05495" (UID: "0aba8976-47dd-4f37-9e4d-e561b7f05495"). InnerVolumeSpecName "kube-api-access-hp6pv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.524803 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7qxc\" (UniqueName: \"kubernetes.io/projected/84053541-a741-4dd3-8b4c-420cfa5da896-kube-api-access-h7qxc\") pod \"84053541-a741-4dd3-8b4c-420cfa5da896\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.525544 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84053541-a741-4dd3-8b4c-420cfa5da896-operator-scripts\") pod \"84053541-a741-4dd3-8b4c-420cfa5da896\" (UID: \"84053541-a741-4dd3-8b4c-420cfa5da896\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.525709 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7dnw\" (UniqueName: \"kubernetes.io/projected/2d8c5715-e9fb-4eef-994f-04aeb6b18864-kube-api-access-h7dnw\") pod \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.525736 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8c5715-e9fb-4eef-994f-04aeb6b18864-operator-scripts\") pod \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\" (UID: \"2d8c5715-e9fb-4eef-994f-04aeb6b18864\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.525765 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvb9t\" (UniqueName: \"kubernetes.io/projected/d59d1326-3f4e-4429-93ee-4b983de7ff82-kube-api-access-zvb9t\") pod \"d59d1326-3f4e-4429-93ee-4b983de7ff82\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.525798 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d59d1326-3f4e-4429-93ee-4b983de7ff82-operator-scripts\") pod \"d59d1326-3f4e-4429-93ee-4b983de7ff82\" (UID: \"d59d1326-3f4e-4429-93ee-4b983de7ff82\") " Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.526163 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp6pv\" (UniqueName: \"kubernetes.io/projected/0aba8976-47dd-4f37-9e4d-e561b7f05495-kube-api-access-hp6pv\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.526178 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5lb5\" (UniqueName: \"kubernetes.io/projected/21c3e515-5917-4360-8ab2-5fc920f8f537-kube-api-access-x5lb5\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.526189 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zftht\" (UniqueName: \"kubernetes.io/projected/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79-kube-api-access-zftht\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.526197 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2q859\" (UniqueName: \"kubernetes.io/projected/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f-kube-api-access-2q859\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.526563 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d59d1326-3f4e-4429-93ee-4b983de7ff82-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d59d1326-3f4e-4429-93ee-4b983de7ff82" (UID: "d59d1326-3f4e-4429-93ee-4b983de7ff82"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.526921 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84053541-a741-4dd3-8b4c-420cfa5da896-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84053541-a741-4dd3-8b4c-420cfa5da896" (UID: "84053541-a741-4dd3-8b4c-420cfa5da896"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.528816 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d8c5715-e9fb-4eef-994f-04aeb6b18864-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d8c5715-e9fb-4eef-994f-04aeb6b18864" (UID: "2d8c5715-e9fb-4eef-994f-04aeb6b18864"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.545591 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84053541-a741-4dd3-8b4c-420cfa5da896-kube-api-access-h7qxc" (OuterVolumeSpecName: "kube-api-access-h7qxc") pod "84053541-a741-4dd3-8b4c-420cfa5da896" (UID: "84053541-a741-4dd3-8b4c-420cfa5da896"). InnerVolumeSpecName "kube-api-access-h7qxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.545635 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8c5715-e9fb-4eef-994f-04aeb6b18864-kube-api-access-h7dnw" (OuterVolumeSpecName: "kube-api-access-h7dnw") pod "2d8c5715-e9fb-4eef-994f-04aeb6b18864" (UID: "2d8c5715-e9fb-4eef-994f-04aeb6b18864"). InnerVolumeSpecName "kube-api-access-h7dnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.547802 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d59d1326-3f4e-4429-93ee-4b983de7ff82-kube-api-access-zvb9t" (OuterVolumeSpecName: "kube-api-access-zvb9t") pod "d59d1326-3f4e-4429-93ee-4b983de7ff82" (UID: "d59d1326-3f4e-4429-93ee-4b983de7ff82"). InnerVolumeSpecName "kube-api-access-zvb9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.629205 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84053541-a741-4dd3-8b4c-420cfa5da896-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.629236 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7dnw\" (UniqueName: \"kubernetes.io/projected/2d8c5715-e9fb-4eef-994f-04aeb6b18864-kube-api-access-h7dnw\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.629248 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d8c5715-e9fb-4eef-994f-04aeb6b18864-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.629256 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvb9t\" (UniqueName: \"kubernetes.io/projected/d59d1326-3f4e-4429-93ee-4b983de7ff82-kube-api-access-zvb9t\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.629269 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d59d1326-3f4e-4429-93ee-4b983de7ff82-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.629278 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7qxc\" (UniqueName: \"kubernetes.io/projected/84053541-a741-4dd3-8b4c-420cfa5da896-kube-api-access-h7qxc\") on node \"crc\" DevicePath \"\"" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.751803 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-qjqbf" event={"ID":"d59d1326-3f4e-4429-93ee-4b983de7ff82","Type":"ContainerDied","Data":"c16f52ddcd21c58660d24a0043acbf8206dab239308436af46b89327755f5ec5"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.751850 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c16f52ddcd21c58660d24a0043acbf8206dab239308436af46b89327755f5ec5" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.751933 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-qjqbf" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.755355 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-91fd-account-create-update-68gbg" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.755346 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-91fd-account-create-update-68gbg" event={"ID":"84053541-a741-4dd3-8b4c-420cfa5da896","Type":"ContainerDied","Data":"f36df9abecf22d72e391e6a9cd10cb832a5ec2d3c8a6497a8a5af5a244022b22"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.755528 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f36df9abecf22d72e391e6a9cd10cb832a5ec2d3c8a6497a8a5af5a244022b22" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.757016 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-c12a-account-create-update-qx47p" event={"ID":"5c067dc9-f4a3-46d3-bb5f-bc8f36312a79","Type":"ContainerDied","Data":"7436484a3fe139c2d26a0eb299621147fbb1af4df6426a8d1dec41730d838d7b"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.757042 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7436484a3fe139c2d26a0eb299621147fbb1af4df6426a8d1dec41730d838d7b" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.757109 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-c12a-account-create-update-qx47p" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.761640 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-5znh9" event={"ID":"21c3e515-5917-4360-8ab2-5fc920f8f537","Type":"ContainerDied","Data":"396af04a6f81eeac1bd82d0c4708d70085d4ac6b59496194e14ca87ab01d6ab3"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.761672 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="396af04a6f81eeac1bd82d0c4708d70085d4ac6b59496194e14ca87ab01d6ab3" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.761696 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-5znh9" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.765812 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-q26vt" event={"ID":"64054bf5-de18-49b3-93c5-e5081ef122f8","Type":"ContainerDied","Data":"cfcdc64de16b2037e2057d5899cea30c519a4dcfdabbe32f25442b36ffc4b038"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.765839 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfcdc64de16b2037e2057d5899cea30c519a4dcfdabbe32f25442b36ffc4b038" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.765894 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-q26vt" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.770185 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-02b5-account-create-update-nlqzc" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.770771 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-02b5-account-create-update-nlqzc" event={"ID":"b1e8d47d-5fa9-4cae-8164-8b5bf089c82f","Type":"ContainerDied","Data":"a01c92528cde8ab08dd2b43c85eb0143294012db957111e583346ce8950eb020"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.770846 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a01c92528cde8ab08dd2b43c85eb0143294012db957111e583346ce8950eb020" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.774686 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lq7fc" event={"ID":"0aba8976-47dd-4f37-9e4d-e561b7f05495","Type":"ContainerDied","Data":"dd1de0a2d4feb29acfc10f005d75f39b4f301d3457fc8535eaed17e402cb972b"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.774765 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd1de0a2d4feb29acfc10f005d75f39b4f301d3457fc8535eaed17e402cb972b" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.774875 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lq7fc" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.779029 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7e35-account-create-update-22fl8" event={"ID":"2d8c5715-e9fb-4eef-994f-04aeb6b18864","Type":"ContainerDied","Data":"c7a35f635e5a7ab346de5c21ec1121bac3936b2544cec3196c0c6d7adec2ab25"} Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.779081 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a35f635e5a7ab346de5c21ec1121bac3936b2544cec3196c0c6d7adec2ab25" Feb 20 06:55:39 crc kubenswrapper[4492]: I0220 06:55:39.779058 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7e35-account-create-update-22fl8" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182010 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rj4bm"] Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182738 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84053541-a741-4dd3-8b4c-420cfa5da896" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182754 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="84053541-a741-4dd3-8b4c-420cfa5da896" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182770 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8c5715-e9fb-4eef-994f-04aeb6b18864" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182776 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8c5715-e9fb-4eef-994f-04aeb6b18864" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182792 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c067dc9-f4a3-46d3-bb5f-bc8f36312a79" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182798 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c067dc9-f4a3-46d3-bb5f-bc8f36312a79" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182811 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d59d1326-3f4e-4429-93ee-4b983de7ff82" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182816 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d59d1326-3f4e-4429-93ee-4b983de7ff82" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182827 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64054bf5-de18-49b3-93c5-e5081ef122f8" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182833 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="64054bf5-de18-49b3-93c5-e5081ef122f8" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182845 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aba8976-47dd-4f37-9e4d-e561b7f05495" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182850 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aba8976-47dd-4f37-9e4d-e561b7f05495" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182864 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21c3e515-5917-4360-8ab2-5fc920f8f537" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182870 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="21c3e515-5917-4360-8ab2-5fc920f8f537" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: E0220 06:55:40.182878 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e8d47d-5fa9-4cae-8164-8b5bf089c82f" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.182883 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e8d47d-5fa9-4cae-8164-8b5bf089c82f" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183059 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="64054bf5-de18-49b3-93c5-e5081ef122f8" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183068 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d59d1326-3f4e-4429-93ee-4b983de7ff82" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183088 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aba8976-47dd-4f37-9e4d-e561b7f05495" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183097 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="84053541-a741-4dd3-8b4c-420cfa5da896" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183106 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8c5715-e9fb-4eef-994f-04aeb6b18864" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183115 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c067dc9-f4a3-46d3-bb5f-bc8f36312a79" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183123 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e8d47d-5fa9-4cae-8164-8b5bf089c82f" containerName="mariadb-account-create-update" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.183129 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="21c3e515-5917-4360-8ab2-5fc920f8f537" containerName="mariadb-database-create" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.184352 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.216022 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj4bm"] Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.239334 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-utilities\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.239434 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8gsx\" (UniqueName: \"kubernetes.io/projected/0a5eca8b-519e-46b4-b2d7-f4c823192a95-kube-api-access-m8gsx\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.239504 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-catalog-content\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.341574 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-utilities\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.341663 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8gsx\" (UniqueName: \"kubernetes.io/projected/0a5eca8b-519e-46b4-b2d7-f4c823192a95-kube-api-access-m8gsx\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.341755 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-catalog-content\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.342329 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-catalog-content\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.344145 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-utilities\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.375462 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8gsx\" (UniqueName: \"kubernetes.io/projected/0a5eca8b-519e-46b4-b2d7-f4c823192a95-kube-api-access-m8gsx\") pod \"certified-operators-rj4bm\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:40 crc kubenswrapper[4492]: I0220 06:55:40.505855 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.744233 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fk2z4"] Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.746253 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.769510 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fk2z4"] Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.895303 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-utilities\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.895647 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsv4z\" (UniqueName: \"kubernetes.io/projected/7047fec7-0f28-4f2f-b655-7f761c01c0f8-kube-api-access-qsv4z\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.895822 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-catalog-content\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.997827 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-utilities\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.997951 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsv4z\" (UniqueName: \"kubernetes.io/projected/7047fec7-0f28-4f2f-b655-7f761c01c0f8-kube-api-access-qsv4z\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.998025 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-catalog-content\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.998370 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-utilities\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:42 crc kubenswrapper[4492]: I0220 06:55:42.998540 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-catalog-content\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:43 crc kubenswrapper[4492]: I0220 06:55:43.028290 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsv4z\" (UniqueName: \"kubernetes.io/projected/7047fec7-0f28-4f2f-b655-7f761c01c0f8-kube-api-access-qsv4z\") pod \"community-operators-fk2z4\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:43 crc kubenswrapper[4492]: I0220 06:55:43.086207 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:55:51 crc kubenswrapper[4492]: E0220 06:55:51.430616 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-keystone:f0473f3e18dd17d7021c02e991298923" Feb 20 06:55:51 crc kubenswrapper[4492]: E0220 06:55:51.431349 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-keystone:f0473f3e18dd17d7021c02e991298923" Feb 20 06:55:51 crc kubenswrapper[4492]: E0220 06:55:51.431510 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-keystone:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m6sjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-x5r2g_openstack(eaa400f5-22b0-44f7-9655-f8b3bd1f8467): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:55:51 crc kubenswrapper[4492]: E0220 06:55:51.432711 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-x5r2g" podUID="eaa400f5-22b0-44f7-9655-f8b3bd1f8467" Feb 20 06:55:51 crc kubenswrapper[4492]: I0220 06:55:51.933336 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fk2z4"] Feb 20 06:55:51 crc kubenswrapper[4492]: E0220 06:55:51.943554 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-keystone:f0473f3e18dd17d7021c02e991298923\\\"\"" pod="openstack/keystone-db-sync-x5r2g" podUID="eaa400f5-22b0-44f7-9655-f8b3bd1f8467" Feb 20 06:55:51 crc kubenswrapper[4492]: I0220 06:55:51.999184 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rj4bm"] Feb 20 06:55:52 crc kubenswrapper[4492]: W0220 06:55:52.011684 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a5eca8b_519e_46b4_b2d7_f4c823192a95.slice/crio-a7a0051e02abaafd315e831491f1a88085b704fdbe8188952ac1292b9915ac79 WatchSource:0}: Error finding container a7a0051e02abaafd315e831491f1a88085b704fdbe8188952ac1292b9915ac79: Status 404 returned error can't find the container with id a7a0051e02abaafd315e831491f1a88085b704fdbe8188952ac1292b9915ac79 Feb 20 06:55:52 crc kubenswrapper[4492]: I0220 06:55:52.953827 4492 generic.go:334] "Generic (PLEG): container finished" podID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerID="94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2" exitCode=0 Feb 20 06:55:52 crc kubenswrapper[4492]: I0220 06:55:52.954190 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj4bm" event={"ID":"0a5eca8b-519e-46b4-b2d7-f4c823192a95","Type":"ContainerDied","Data":"94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2"} Feb 20 06:55:52 crc kubenswrapper[4492]: I0220 06:55:52.954230 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj4bm" event={"ID":"0a5eca8b-519e-46b4-b2d7-f4c823192a95","Type":"ContainerStarted","Data":"a7a0051e02abaafd315e831491f1a88085b704fdbe8188952ac1292b9915ac79"} Feb 20 06:55:52 crc kubenswrapper[4492]: I0220 06:55:52.960205 4492 generic.go:334] "Generic (PLEG): container finished" podID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerID="743ff5953bbff4c6e4305c24bb0aac565b11b9ad87e915ec759acc6020da3410" exitCode=0 Feb 20 06:55:52 crc kubenswrapper[4492]: I0220 06:55:52.960296 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fk2z4" event={"ID":"7047fec7-0f28-4f2f-b655-7f761c01c0f8","Type":"ContainerDied","Data":"743ff5953bbff4c6e4305c24bb0aac565b11b9ad87e915ec759acc6020da3410"} Feb 20 06:55:52 crc kubenswrapper[4492]: I0220 06:55:52.960337 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fk2z4" event={"ID":"7047fec7-0f28-4f2f-b655-7f761c01c0f8","Type":"ContainerStarted","Data":"1cf88cec665d9f24cca6bd7859340b6a1da869dbff784402a1bb78e66afbe349"} Feb 20 06:55:52 crc kubenswrapper[4492]: I0220 06:55:52.963515 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scs24" event={"ID":"8e4ff2bf-0f92-4271-bc11-e07604d93c5b","Type":"ContainerStarted","Data":"3ce12d434ad1ecd2f28b179736935ad1433b0d8a589f939bd4357a21ff1e4f6a"} Feb 20 06:55:53 crc kubenswrapper[4492]: I0220 06:55:53.025633 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-scs24" podStartSLOduration=3.2981657540000002 podStartE2EDuration="36.025602449s" podCreationTimestamp="2026-02-20 06:55:17 +0000 UTC" firstStartedPulling="2026-02-20 06:55:18.800255938 +0000 UTC m=+875.571544916" lastFinishedPulling="2026-02-20 06:55:51.527692633 +0000 UTC m=+908.298981611" observedRunningTime="2026-02-20 06:55:53.023221621 +0000 UTC m=+909.794510599" watchObservedRunningTime="2026-02-20 06:55:53.025602449 +0000 UTC m=+909.796891426" Feb 20 06:55:53 crc kubenswrapper[4492]: I0220 06:55:53.976173 4492 generic.go:334] "Generic (PLEG): container finished" podID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerID="89787f2e7b36a1c256cfe2afc5defd2dbb6abf029215bbd3c5c0fadde37b53f5" exitCode=0 Feb 20 06:55:53 crc kubenswrapper[4492]: I0220 06:55:53.976216 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fk2z4" event={"ID":"7047fec7-0f28-4f2f-b655-7f761c01c0f8","Type":"ContainerDied","Data":"89787f2e7b36a1c256cfe2afc5defd2dbb6abf029215bbd3c5c0fadde37b53f5"} Feb 20 06:55:54 crc kubenswrapper[4492]: I0220 06:55:54.006838 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj4bm" event={"ID":"0a5eca8b-519e-46b4-b2d7-f4c823192a95","Type":"ContainerStarted","Data":"35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f"} Feb 20 06:55:55 crc kubenswrapper[4492]: I0220 06:55:55.014791 4492 generic.go:334] "Generic (PLEG): container finished" podID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerID="35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f" exitCode=0 Feb 20 06:55:55 crc kubenswrapper[4492]: I0220 06:55:55.015007 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj4bm" event={"ID":"0a5eca8b-519e-46b4-b2d7-f4c823192a95","Type":"ContainerDied","Data":"35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f"} Feb 20 06:55:55 crc kubenswrapper[4492]: I0220 06:55:55.017606 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fk2z4" event={"ID":"7047fec7-0f28-4f2f-b655-7f761c01c0f8","Type":"ContainerStarted","Data":"d4994787df3d8eee9f508b0337c9c42001058e1cdef99ce73b1f2ddee6441211"} Feb 20 06:55:56 crc kubenswrapper[4492]: I0220 06:55:56.031125 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj4bm" event={"ID":"0a5eca8b-519e-46b4-b2d7-f4c823192a95","Type":"ContainerStarted","Data":"b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324"} Feb 20 06:55:56 crc kubenswrapper[4492]: I0220 06:55:56.056528 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rj4bm" podStartSLOduration=13.510468248 podStartE2EDuration="16.056457876s" podCreationTimestamp="2026-02-20 06:55:40 +0000 UTC" firstStartedPulling="2026-02-20 06:55:52.956591734 +0000 UTC m=+909.727880712" lastFinishedPulling="2026-02-20 06:55:55.502581362 +0000 UTC m=+912.273870340" observedRunningTime="2026-02-20 06:55:56.055994722 +0000 UTC m=+912.827283700" watchObservedRunningTime="2026-02-20 06:55:56.056457876 +0000 UTC m=+912.827746853" Feb 20 06:55:56 crc kubenswrapper[4492]: I0220 06:55:56.062678 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fk2z4" podStartSLOduration=12.483179899 podStartE2EDuration="14.062661397s" podCreationTimestamp="2026-02-20 06:55:42 +0000 UTC" firstStartedPulling="2026-02-20 06:55:52.962629814 +0000 UTC m=+909.733918791" lastFinishedPulling="2026-02-20 06:55:54.54211131 +0000 UTC m=+911.313400289" observedRunningTime="2026-02-20 06:55:55.052832218 +0000 UTC m=+911.824121195" watchObservedRunningTime="2026-02-20 06:55:56.062661397 +0000 UTC m=+912.833950375" Feb 20 06:55:59 crc kubenswrapper[4492]: I0220 06:55:59.060145 4492 generic.go:334] "Generic (PLEG): container finished" podID="8e4ff2bf-0f92-4271-bc11-e07604d93c5b" containerID="3ce12d434ad1ecd2f28b179736935ad1433b0d8a589f939bd4357a21ff1e4f6a" exitCode=0 Feb 20 06:55:59 crc kubenswrapper[4492]: I0220 06:55:59.060249 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scs24" event={"ID":"8e4ff2bf-0f92-4271-bc11-e07604d93c5b","Type":"ContainerDied","Data":"3ce12d434ad1ecd2f28b179736935ad1433b0d8a589f939bd4357a21ff1e4f6a"} Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.156460 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z7d95"] Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.158887 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.163582 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-catalog-content\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.163636 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-utilities\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.163711 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28cvr\" (UniqueName: \"kubernetes.io/projected/6617a084-7a2b-4b2a-bec2-d1eb4378c841-kube-api-access-28cvr\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.164226 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7d95"] Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.266074 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-catalog-content\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.266113 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-utilities\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.266186 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28cvr\" (UniqueName: \"kubernetes.io/projected/6617a084-7a2b-4b2a-bec2-d1eb4378c841-kube-api-access-28cvr\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.266683 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-catalog-content\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.266687 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-utilities\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.293643 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28cvr\" (UniqueName: \"kubernetes.io/projected/6617a084-7a2b-4b2a-bec2-d1eb4378c841-kube-api-access-28cvr\") pod \"redhat-marketplace-z7d95\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.475430 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.507322 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.507468 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.550361 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.577163 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scs24" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.674442 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-db-sync-config-data\") pod \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.674511 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-config-data\") pod \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.674649 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-combined-ca-bundle\") pod \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.674924 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctlgg\" (UniqueName: \"kubernetes.io/projected/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-kube-api-access-ctlgg\") pod \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\" (UID: \"8e4ff2bf-0f92-4271-bc11-e07604d93c5b\") " Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.685696 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8e4ff2bf-0f92-4271-bc11-e07604d93c5b" (UID: "8e4ff2bf-0f92-4271-bc11-e07604d93c5b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.689874 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-kube-api-access-ctlgg" (OuterVolumeSpecName: "kube-api-access-ctlgg") pod "8e4ff2bf-0f92-4271-bc11-e07604d93c5b" (UID: "8e4ff2bf-0f92-4271-bc11-e07604d93c5b"). InnerVolumeSpecName "kube-api-access-ctlgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.714772 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e4ff2bf-0f92-4271-bc11-e07604d93c5b" (UID: "8e4ff2bf-0f92-4271-bc11-e07604d93c5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.775576 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-config-data" (OuterVolumeSpecName: "config-data") pod "8e4ff2bf-0f92-4271-bc11-e07604d93c5b" (UID: "8e4ff2bf-0f92-4271-bc11-e07604d93c5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.777982 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctlgg\" (UniqueName: \"kubernetes.io/projected/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-kube-api-access-ctlgg\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.778011 4492 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.778022 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:00 crc kubenswrapper[4492]: I0220 06:56:00.778034 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e4ff2bf-0f92-4271-bc11-e07604d93c5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.015014 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7d95"] Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.085258 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-scs24" event={"ID":"8e4ff2bf-0f92-4271-bc11-e07604d93c5b","Type":"ContainerDied","Data":"16235fca0bf2919e8ef631200bfa6d8bd998d5a64cc5e5add2719bfe9176d1a7"} Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.085323 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16235fca0bf2919e8ef631200bfa6d8bd998d5a64cc5e5add2719bfe9176d1a7" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.085462 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-scs24" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.105492 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7d95" event={"ID":"6617a084-7a2b-4b2a-bec2-d1eb4378c841","Type":"ContainerStarted","Data":"fa0d81279a8206274ca8aa4cc48eec6dda96e116396044da30993789d4e8ab90"} Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.180689 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.507598 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8668cb6c6f-2bknz"] Feb 20 06:56:01 crc kubenswrapper[4492]: E0220 06:56:01.507975 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e4ff2bf-0f92-4271-bc11-e07604d93c5b" containerName="glance-db-sync" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.507995 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e4ff2bf-0f92-4271-bc11-e07604d93c5b" containerName="glance-db-sync" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.508169 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e4ff2bf-0f92-4271-bc11-e07604d93c5b" containerName="glance-db-sync" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.508984 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.534085 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8668cb6c6f-2bknz"] Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.594031 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-config\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.594168 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-svc\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.594189 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-nb\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.594258 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkj9d\" (UniqueName: \"kubernetes.io/projected/cca7a0af-f44b-43f8-8f45-894ee3609a0a-kube-api-access-nkj9d\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.594304 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-swift-storage-0\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.594517 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-sb\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.696533 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-swift-storage-0\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.696972 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-sb\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.697130 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-config\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.697306 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-svc\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.697330 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-nb\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.697388 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkj9d\" (UniqueName: \"kubernetes.io/projected/cca7a0af-f44b-43f8-8f45-894ee3609a0a-kube-api-access-nkj9d\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.697547 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-swift-storage-0\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.698060 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-sb\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.698244 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-config\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.698444 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-svc\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.698682 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-nb\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.727253 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkj9d\" (UniqueName: \"kubernetes.io/projected/cca7a0af-f44b-43f8-8f45-894ee3609a0a-kube-api-access-nkj9d\") pod \"dnsmasq-dns-8668cb6c6f-2bknz\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:01 crc kubenswrapper[4492]: I0220 06:56:01.823298 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:02 crc kubenswrapper[4492]: I0220 06:56:02.119761 4492 generic.go:334] "Generic (PLEG): container finished" podID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerID="0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7" exitCode=0 Feb 20 06:56:02 crc kubenswrapper[4492]: I0220 06:56:02.121035 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7d95" event={"ID":"6617a084-7a2b-4b2a-bec2-d1eb4378c841","Type":"ContainerDied","Data":"0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7"} Feb 20 06:56:02 crc kubenswrapper[4492]: I0220 06:56:02.339816 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8668cb6c6f-2bknz"] Feb 20 06:56:02 crc kubenswrapper[4492]: W0220 06:56:02.353682 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcca7a0af_f44b_43f8_8f45_894ee3609a0a.slice/crio-b96bf847e3e2b503106bbe09e88d7307af6bbaa24054d01881748bce880070fc WatchSource:0}: Error finding container b96bf847e3e2b503106bbe09e88d7307af6bbaa24054d01881748bce880070fc: Status 404 returned error can't find the container with id b96bf847e3e2b503106bbe09e88d7307af6bbaa24054d01881748bce880070fc Feb 20 06:56:02 crc kubenswrapper[4492]: I0220 06:56:02.945198 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj4bm"] Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.087129 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.087511 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.126130 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.128227 4492 generic.go:334] "Generic (PLEG): container finished" podID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerID="e9c3e15217b9ab793bffd909053306530a37c43aad80d26d3218df2ed4c849bf" exitCode=0 Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.128331 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" event={"ID":"cca7a0af-f44b-43f8-8f45-894ee3609a0a","Type":"ContainerDied","Data":"e9c3e15217b9ab793bffd909053306530a37c43aad80d26d3218df2ed4c849bf"} Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.128399 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" event={"ID":"cca7a0af-f44b-43f8-8f45-894ee3609a0a","Type":"ContainerStarted","Data":"b96bf847e3e2b503106bbe09e88d7307af6bbaa24054d01881748bce880070fc"} Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.131794 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7d95" event={"ID":"6617a084-7a2b-4b2a-bec2-d1eb4378c841","Type":"ContainerStarted","Data":"a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9"} Feb 20 06:56:03 crc kubenswrapper[4492]: I0220 06:56:03.174653 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.158217 4492 generic.go:334] "Generic (PLEG): container finished" podID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerID="a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9" exitCode=0 Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.158666 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7d95" event={"ID":"6617a084-7a2b-4b2a-bec2-d1eb4378c841","Type":"ContainerDied","Data":"a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9"} Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.162099 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" event={"ID":"cca7a0af-f44b-43f8-8f45-894ee3609a0a","Type":"ContainerStarted","Data":"f7ee6c3fee157ec60a72a08f540ad9f3e5137b84d114ba4079b7b7b27552668f"} Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.162955 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rj4bm" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="registry-server" containerID="cri-o://b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324" gracePeriod=2 Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.191806 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" podStartSLOduration=3.191787633 podStartE2EDuration="3.191787633s" podCreationTimestamp="2026-02-20 06:56:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:04.18979561 +0000 UTC m=+920.961084587" watchObservedRunningTime="2026-02-20 06:56:04.191787633 +0000 UTC m=+920.963076611" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.648720 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.771769 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-catalog-content\") pod \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.772273 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8gsx\" (UniqueName: \"kubernetes.io/projected/0a5eca8b-519e-46b4-b2d7-f4c823192a95-kube-api-access-m8gsx\") pod \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.772329 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-utilities\") pod \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\" (UID: \"0a5eca8b-519e-46b4-b2d7-f4c823192a95\") " Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.773584 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-utilities" (OuterVolumeSpecName: "utilities") pod "0a5eca8b-519e-46b4-b2d7-f4c823192a95" (UID: "0a5eca8b-519e-46b4-b2d7-f4c823192a95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.780345 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a5eca8b-519e-46b4-b2d7-f4c823192a95-kube-api-access-m8gsx" (OuterVolumeSpecName: "kube-api-access-m8gsx") pod "0a5eca8b-519e-46b4-b2d7-f4c823192a95" (UID: "0a5eca8b-519e-46b4-b2d7-f4c823192a95"). InnerVolumeSpecName "kube-api-access-m8gsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.832223 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a5eca8b-519e-46b4-b2d7-f4c823192a95" (UID: "0a5eca8b-519e-46b4-b2d7-f4c823192a95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.874958 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8gsx\" (UniqueName: \"kubernetes.io/projected/0a5eca8b-519e-46b4-b2d7-f4c823192a95-kube-api-access-m8gsx\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.874992 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:04 crc kubenswrapper[4492]: I0220 06:56:04.875004 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a5eca8b-519e-46b4-b2d7-f4c823192a95-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.174873 4492 generic.go:334] "Generic (PLEG): container finished" podID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerID="b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324" exitCode=0 Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.174953 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj4bm" event={"ID":"0a5eca8b-519e-46b4-b2d7-f4c823192a95","Type":"ContainerDied","Data":"b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324"} Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.174966 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rj4bm" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.174992 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rj4bm" event={"ID":"0a5eca8b-519e-46b4-b2d7-f4c823192a95","Type":"ContainerDied","Data":"a7a0051e02abaafd315e831491f1a88085b704fdbe8188952ac1292b9915ac79"} Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.175016 4492 scope.go:117] "RemoveContainer" containerID="b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.183559 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7d95" event={"ID":"6617a084-7a2b-4b2a-bec2-d1eb4378c841","Type":"ContainerStarted","Data":"a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097"} Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.183588 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.207750 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z7d95" podStartSLOduration=2.70137654 podStartE2EDuration="5.20773857s" podCreationTimestamp="2026-02-20 06:56:00 +0000 UTC" firstStartedPulling="2026-02-20 06:56:02.122308598 +0000 UTC m=+918.893597576" lastFinishedPulling="2026-02-20 06:56:04.628670628 +0000 UTC m=+921.399959606" observedRunningTime="2026-02-20 06:56:05.200789132 +0000 UTC m=+921.972078111" watchObservedRunningTime="2026-02-20 06:56:05.20773857 +0000 UTC m=+921.979027548" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.208677 4492 scope.go:117] "RemoveContainer" containerID="35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.219342 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rj4bm"] Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.227254 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rj4bm"] Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.241603 4492 scope.go:117] "RemoveContainer" containerID="94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.264782 4492 scope.go:117] "RemoveContainer" containerID="b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324" Feb 20 06:56:05 crc kubenswrapper[4492]: E0220 06:56:05.265728 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324\": container with ID starting with b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324 not found: ID does not exist" containerID="b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.265817 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324"} err="failed to get container status \"b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324\": rpc error: code = NotFound desc = could not find container \"b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324\": container with ID starting with b06297080cea49a8650bad1799ea1bf8fca32535b1b5fbd3aee4052b65078324 not found: ID does not exist" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.265895 4492 scope.go:117] "RemoveContainer" containerID="35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f" Feb 20 06:56:05 crc kubenswrapper[4492]: E0220 06:56:05.266513 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f\": container with ID starting with 35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f not found: ID does not exist" containerID="35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.266569 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f"} err="failed to get container status \"35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f\": rpc error: code = NotFound desc = could not find container \"35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f\": container with ID starting with 35067ac51325b7c5c8f8aa6798a5dd07d40bbc9103e8309c333a1ddfc6270e9f not found: ID does not exist" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.266607 4492 scope.go:117] "RemoveContainer" containerID="94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2" Feb 20 06:56:05 crc kubenswrapper[4492]: E0220 06:56:05.268943 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2\": container with ID starting with 94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2 not found: ID does not exist" containerID="94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.269047 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2"} err="failed to get container status \"94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2\": rpc error: code = NotFound desc = could not find container \"94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2\": container with ID starting with 94e404784905f16cd5ac667e4a1965d5551e78a046661af7bead08be03f998c2 not found: ID does not exist" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.570410 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" path="/var/lib/kubelet/pods/0a5eca8b-519e-46b4-b2d7-f4c823192a95/volumes" Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.945990 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fk2z4"] Feb 20 06:56:05 crc kubenswrapper[4492]: I0220 06:56:05.946191 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fk2z4" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="registry-server" containerID="cri-o://d4994787df3d8eee9f508b0337c9c42001058e1cdef99ce73b1f2ddee6441211" gracePeriod=2 Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.197153 4492 generic.go:334] "Generic (PLEG): container finished" podID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerID="d4994787df3d8eee9f508b0337c9c42001058e1cdef99ce73b1f2ddee6441211" exitCode=0 Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.197456 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fk2z4" event={"ID":"7047fec7-0f28-4f2f-b655-7f761c01c0f8","Type":"ContainerDied","Data":"d4994787df3d8eee9f508b0337c9c42001058e1cdef99ce73b1f2ddee6441211"} Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.414810 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.610309 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsv4z\" (UniqueName: \"kubernetes.io/projected/7047fec7-0f28-4f2f-b655-7f761c01c0f8-kube-api-access-qsv4z\") pod \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.610404 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-utilities\") pod \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.610518 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-catalog-content\") pod \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\" (UID: \"7047fec7-0f28-4f2f-b655-7f761c01c0f8\") " Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.611085 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-utilities" (OuterVolumeSpecName: "utilities") pod "7047fec7-0f28-4f2f-b655-7f761c01c0f8" (UID: "7047fec7-0f28-4f2f-b655-7f761c01c0f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.616329 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7047fec7-0f28-4f2f-b655-7f761c01c0f8-kube-api-access-qsv4z" (OuterVolumeSpecName: "kube-api-access-qsv4z") pod "7047fec7-0f28-4f2f-b655-7f761c01c0f8" (UID: "7047fec7-0f28-4f2f-b655-7f761c01c0f8"). InnerVolumeSpecName "kube-api-access-qsv4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.654870 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7047fec7-0f28-4f2f-b655-7f761c01c0f8" (UID: "7047fec7-0f28-4f2f-b655-7f761c01c0f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.713577 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsv4z\" (UniqueName: \"kubernetes.io/projected/7047fec7-0f28-4f2f-b655-7f761c01c0f8-kube-api-access-qsv4z\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.713612 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:06 crc kubenswrapper[4492]: I0220 06:56:06.713632 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7047fec7-0f28-4f2f-b655-7f761c01c0f8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.209259 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fk2z4" event={"ID":"7047fec7-0f28-4f2f-b655-7f761c01c0f8","Type":"ContainerDied","Data":"1cf88cec665d9f24cca6bd7859340b6a1da869dbff784402a1bb78e66afbe349"} Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.209318 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fk2z4" Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.209333 4492 scope.go:117] "RemoveContainer" containerID="d4994787df3d8eee9f508b0337c9c42001058e1cdef99ce73b1f2ddee6441211" Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.211529 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x5r2g" event={"ID":"eaa400f5-22b0-44f7-9655-f8b3bd1f8467","Type":"ContainerStarted","Data":"d40bf6360c8b922955e98ea57d1c630a3d16f148de5c5d4a4e824041c59ebf58"} Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.236223 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-x5r2g" podStartSLOduration=2.690138889 podStartE2EDuration="32.236206472s" podCreationTimestamp="2026-02-20 06:55:35 +0000 UTC" firstStartedPulling="2026-02-20 06:55:36.678055662 +0000 UTC m=+893.449344640" lastFinishedPulling="2026-02-20 06:56:06.224123244 +0000 UTC m=+922.995412223" observedRunningTime="2026-02-20 06:56:07.233356129 +0000 UTC m=+924.004645107" watchObservedRunningTime="2026-02-20 06:56:07.236206472 +0000 UTC m=+924.007495449" Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.237239 4492 scope.go:117] "RemoveContainer" containerID="89787f2e7b36a1c256cfe2afc5defd2dbb6abf029215bbd3c5c0fadde37b53f5" Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.256804 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fk2z4"] Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.262851 4492 scope.go:117] "RemoveContainer" containerID="743ff5953bbff4c6e4305c24bb0aac565b11b9ad87e915ec759acc6020da3410" Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.263107 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fk2z4"] Feb 20 06:56:07 crc kubenswrapper[4492]: I0220 06:56:07.567310 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" path="/var/lib/kubelet/pods/7047fec7-0f28-4f2f-b655-7f761c01c0f8/volumes" Feb 20 06:56:09 crc kubenswrapper[4492]: I0220 06:56:09.230143 4492 generic.go:334] "Generic (PLEG): container finished" podID="eaa400f5-22b0-44f7-9655-f8b3bd1f8467" containerID="d40bf6360c8b922955e98ea57d1c630a3d16f148de5c5d4a4e824041c59ebf58" exitCode=0 Feb 20 06:56:09 crc kubenswrapper[4492]: I0220 06:56:09.230218 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x5r2g" event={"ID":"eaa400f5-22b0-44f7-9655-f8b3bd1f8467","Type":"ContainerDied","Data":"d40bf6360c8b922955e98ea57d1c630a3d16f148de5c5d4a4e824041c59ebf58"} Feb 20 06:56:09 crc kubenswrapper[4492]: I0220 06:56:09.311860 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:56:09 crc kubenswrapper[4492]: I0220 06:56:09.312066 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.475898 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.477267 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.521043 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.550033 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.688249 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-config-data\") pod \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.688395 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-combined-ca-bundle\") pod \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.688955 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6sjs\" (UniqueName: \"kubernetes.io/projected/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-kube-api-access-m6sjs\") pod \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\" (UID: \"eaa400f5-22b0-44f7-9655-f8b3bd1f8467\") " Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.697348 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-kube-api-access-m6sjs" (OuterVolumeSpecName: "kube-api-access-m6sjs") pod "eaa400f5-22b0-44f7-9655-f8b3bd1f8467" (UID: "eaa400f5-22b0-44f7-9655-f8b3bd1f8467"). InnerVolumeSpecName "kube-api-access-m6sjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.710193 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaa400f5-22b0-44f7-9655-f8b3bd1f8467" (UID: "eaa400f5-22b0-44f7-9655-f8b3bd1f8467"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.729305 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-config-data" (OuterVolumeSpecName: "config-data") pod "eaa400f5-22b0-44f7-9655-f8b3bd1f8467" (UID: "eaa400f5-22b0-44f7-9655-f8b3bd1f8467"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.792064 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.792095 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:10 crc kubenswrapper[4492]: I0220 06:56:10.792106 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6sjs\" (UniqueName: \"kubernetes.io/projected/eaa400f5-22b0-44f7-9655-f8b3bd1f8467-kube-api-access-m6sjs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.251265 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-x5r2g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.251779 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-x5r2g" event={"ID":"eaa400f5-22b0-44f7-9655-f8b3bd1f8467","Type":"ContainerDied","Data":"351b1eddb2d6627da07875ebda350c47cd6cbb6f31d9332ad1628c681738b177"} Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.251861 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="351b1eddb2d6627da07875ebda350c47cd6cbb6f31d9332ad1628c681738b177" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.308843 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.502850 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8668cb6c6f-2bknz"] Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.503413 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerName="dnsmasq-dns" containerID="cri-o://f7ee6c3fee157ec60a72a08f540ad9f3e5137b84d114ba4079b7b7b27552668f" gracePeriod=10 Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.507994 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553306 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7495f97fdc-6pbrw"] Feb 20 06:56:11 crc kubenswrapper[4492]: E0220 06:56:11.553816 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="registry-server" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553839 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="registry-server" Feb 20 06:56:11 crc kubenswrapper[4492]: E0220 06:56:11.553858 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="extract-utilities" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553865 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="extract-utilities" Feb 20 06:56:11 crc kubenswrapper[4492]: E0220 06:56:11.553879 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="extract-utilities" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553886 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="extract-utilities" Feb 20 06:56:11 crc kubenswrapper[4492]: E0220 06:56:11.553896 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="extract-content" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553902 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="extract-content" Feb 20 06:56:11 crc kubenswrapper[4492]: E0220 06:56:11.553925 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="extract-content" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553932 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="extract-content" Feb 20 06:56:11 crc kubenswrapper[4492]: E0220 06:56:11.553939 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaa400f5-22b0-44f7-9655-f8b3bd1f8467" containerName="keystone-db-sync" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553944 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaa400f5-22b0-44f7-9655-f8b3bd1f8467" containerName="keystone-db-sync" Feb 20 06:56:11 crc kubenswrapper[4492]: E0220 06:56:11.553958 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="registry-server" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.553962 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="registry-server" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.554166 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7047fec7-0f28-4f2f-b655-7f761c01c0f8" containerName="registry-server" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.554187 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a5eca8b-519e-46b4-b2d7-f4c823192a95" containerName="registry-server" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.554208 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaa400f5-22b0-44f7-9655-f8b3bd1f8467" containerName="keystone-db-sync" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.555108 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.573511 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-h6p7g"] Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.585288 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.592641 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.592757 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4v2r8" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.592933 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.593043 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.598643 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.603133 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7495f97fdc-6pbrw"] Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.643844 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-h6p7g"] Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711328 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-svc\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711394 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-config-data\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711489 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-config\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711533 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-sb\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711566 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-combined-ca-bundle\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711584 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-fernet-keys\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711601 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-nb\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711694 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-scripts\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711723 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-credential-keys\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711764 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrclr\" (UniqueName: \"kubernetes.io/projected/870fda83-5faf-49c8-8b92-4ac16a116677-kube-api-access-rrclr\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711787 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwvm6\" (UniqueName: \"kubernetes.io/projected/9aa62fa4-cbb7-4a89-8de1-91eac6703146-kube-api-access-dwvm6\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.711815 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-swift-storage-0\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813073 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-sb\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813135 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-combined-ca-bundle\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813155 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-fernet-keys\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813175 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-nb\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813234 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-scripts\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813261 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-credential-keys\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813295 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrclr\" (UniqueName: \"kubernetes.io/projected/870fda83-5faf-49c8-8b92-4ac16a116677-kube-api-access-rrclr\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813315 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwvm6\" (UniqueName: \"kubernetes.io/projected/9aa62fa4-cbb7-4a89-8de1-91eac6703146-kube-api-access-dwvm6\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813345 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-swift-storage-0\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813368 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-svc\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813398 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-config-data\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.813462 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-config\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.814281 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-config\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.814822 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-sb\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.816298 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-nb\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.819333 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-swift-storage-0\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.820396 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-combined-ca-bundle\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.820626 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-svc\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.821957 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-fernet-keys\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.829871 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.134:5353: connect: connection refused" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.831397 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-config-data\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.831881 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-credential-keys\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.837228 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7d95"] Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.841968 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-scripts\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.886405 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrclr\" (UniqueName: \"kubernetes.io/projected/870fda83-5faf-49c8-8b92-4ac16a116677-kube-api-access-rrclr\") pod \"dnsmasq-dns-7495f97fdc-6pbrw\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.889875 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.956554 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwvm6\" (UniqueName: \"kubernetes.io/projected/9aa62fa4-cbb7-4a89-8de1-91eac6703146-kube-api-access-dwvm6\") pod \"keystone-bootstrap-h6p7g\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.958287 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-qqhvc"] Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.959448 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.962560 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-s246x"] Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.963435 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-s246x" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.976940 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bj5qz" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.977253 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.977385 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.977394 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 20 06:56:11 crc kubenswrapper[4492]: I0220 06:56:11.977595 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-5wknk" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.011538 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-s246x"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.026272 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6554ff645f-r4p2z"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.027431 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.038702 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-qqhvc"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.051912 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6554ff645f-r4p2z"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.059946 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.060265 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.060414 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.060646 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-rc84t" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.137976 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-config-data\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138255 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2v56\" (UniqueName: \"kubernetes.io/projected/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-kube-api-access-p2v56\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138350 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-config-data\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138373 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-scripts\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138388 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-db-sync-config-data\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138409 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-logs\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138467 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-horizon-secret-key\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138522 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-combined-ca-bundle\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138578 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr62s\" (UniqueName: \"kubernetes.io/projected/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-kube-api-access-dr62s\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138659 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-config-data\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138728 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-combined-ca-bundle\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138750 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-scripts\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.138802 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ed151d-241b-465d-8b5d-d97fe51015df-etc-machine-id\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.145494 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxg4m\" (UniqueName: \"kubernetes.io/projected/82ed151d-241b-465d-8b5d-d97fe51015df-kube-api-access-vxg4m\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.204002 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.240769 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2s268"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.258878 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ed151d-241b-465d-8b5d-d97fe51015df-etc-machine-id\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.258915 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxg4m\" (UniqueName: \"kubernetes.io/projected/82ed151d-241b-465d-8b5d-d97fe51015df-kube-api-access-vxg4m\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.258970 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-config-data\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259013 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2v56\" (UniqueName: \"kubernetes.io/projected/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-kube-api-access-p2v56\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259060 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-scripts\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259075 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-db-sync-config-data\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259092 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-config-data\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259108 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-logs\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259141 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-horizon-secret-key\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259164 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-combined-ca-bundle\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259197 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr62s\" (UniqueName: \"kubernetes.io/projected/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-kube-api-access-dr62s\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259235 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-config-data\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259274 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-combined-ca-bundle\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.259289 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-scripts\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.260098 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.261556 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-scripts\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.262623 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-config-data\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.264907 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-logs\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.264967 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ed151d-241b-465d-8b5d-d97fe51015df-etc-machine-id\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.271911 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-config-data\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.276829 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-combined-ca-bundle\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.277365 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-db-sync-config-data\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.280655 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-scripts\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.291245 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-horizon-secret-key\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.291739 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.292303 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-combined-ca-bundle\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.292623 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.292759 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ndkd9" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.299067 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-config-data\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.321504 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.328116 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.329837 4492 generic.go:334] "Generic (PLEG): container finished" podID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerID="f7ee6c3fee157ec60a72a08f540ad9f3e5137b84d114ba4079b7b7b27552668f" exitCode=0 Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.354625 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2s268"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.354661 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" event={"ID":"cca7a0af-f44b-43f8-8f45-894ee3609a0a","Type":"ContainerDied","Data":"f7ee6c3fee157ec60a72a08f540ad9f3e5137b84d114ba4079b7b7b27552668f"} Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.358940 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.359183 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.366695 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxg4m\" (UniqueName: \"kubernetes.io/projected/82ed151d-241b-465d-8b5d-d97fe51015df-kube-api-access-vxg4m\") pod \"cinder-db-sync-qqhvc\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.367682 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-config\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.367787 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c6fj\" (UniqueName: \"kubernetes.io/projected/4d697f3d-7b65-4983-8602-de839d5caa82-kube-api-access-2c6fj\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.367841 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-combined-ca-bundle\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.369499 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2v56\" (UniqueName: \"kubernetes.io/projected/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-kube-api-access-p2v56\") pod \"heat-db-sync-s246x\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.373941 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr62s\" (UniqueName: \"kubernetes.io/projected/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-kube-api-access-dr62s\") pod \"horizon-6554ff645f-r4p2z\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.394579 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.446857 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5d5c9766b9-5tjrb"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.448677 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.461540 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-d79gm"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.495892 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.516320 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fslbj" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.517067 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.556862 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb8wp\" (UniqueName: \"kubernetes.io/projected/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-kube-api-access-gb8wp\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.556953 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-scripts\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557041 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c6fj\" (UniqueName: \"kubernetes.io/projected/4d697f3d-7b65-4983-8602-de839d5caa82-kube-api-access-2c6fj\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557096 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557132 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-scripts\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557195 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-combined-ca-bundle\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557307 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-config-data\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557436 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-config-data\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557458 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-run-httpd\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557522 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b83b1978-9929-413b-a7b2-6f9bacf9af02-horizon-secret-key\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557567 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557628 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83b1978-9929-413b-a7b2-6f9bacf9af02-logs\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557656 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6v9s\" (UniqueName: \"kubernetes.io/projected/b83b1978-9929-413b-a7b2-6f9bacf9af02-kube-api-access-g6v9s\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557720 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-config\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.557746 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-log-httpd\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.586119 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-config\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.587205 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-combined-ca-bundle\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.597861 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d5c9766b9-5tjrb"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.628692 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.629145 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-s246x" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.639004 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c6fj\" (UniqueName: \"kubernetes.io/projected/4d697f3d-7b65-4983-8602-de839d5caa82-kube-api-access-2c6fj\") pod \"neutron-db-sync-2s268\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.700410 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704066 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-config-data\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704102 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-run-httpd\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704138 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b83b1978-9929-413b-a7b2-6f9bacf9af02-horizon-secret-key\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704179 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704204 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83b1978-9929-413b-a7b2-6f9bacf9af02-logs\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704221 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6v9s\" (UniqueName: \"kubernetes.io/projected/b83b1978-9929-413b-a7b2-6f9bacf9af02-kube-api-access-g6v9s\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704269 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-log-httpd\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704334 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb8wp\" (UniqueName: \"kubernetes.io/projected/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-kube-api-access-gb8wp\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704353 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-scripts\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704403 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.704423 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-scripts\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.705954 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-d79gm"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.719379 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-config-data\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.727287 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gjtqr"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.729506 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.749859 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-log-httpd\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.750043 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-run-httpd\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.750208 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83b1978-9929-413b-a7b2-6f9bacf9af02-logs\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.750631 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.750822 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rr7tt" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.750942 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.756214 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-scripts\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.759901 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-config-data\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.760273 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-scripts\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.763077 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.766228 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.777523 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-config-data\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.777767 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b83b1978-9929-413b-a7b2-6f9bacf9af02-horizon-secret-key\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.834206 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.835525 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z57n\" (UniqueName: \"kubernetes.io/projected/05565eb0-90db-4ec7-b646-db528fca40ad-kube-api-access-5z57n\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.835678 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-combined-ca-bundle\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.835800 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b719757-21ff-4967-a478-287c671d2bc7-logs\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.835857 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-config-data\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.835887 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fllcl\" (UniqueName: \"kubernetes.io/projected/5b719757-21ff-4967-a478-287c671d2bc7-kube-api-access-fllcl\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.835909 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-scripts\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.835991 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-combined-ca-bundle\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.836057 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-db-sync-config-data\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.866293 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb8wp\" (UniqueName: \"kubernetes.io/projected/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-kube-api-access-gb8wp\") pod \"ceilometer-0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.869320 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gjtqr"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.869453 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.872822 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6v9s\" (UniqueName: \"kubernetes.io/projected/b83b1978-9929-413b-a7b2-6f9bacf9af02-kube-api-access-g6v9s\") pod \"horizon-5d5c9766b9-5tjrb\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.876117 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.879687 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jvgvq" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.879866 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.885452 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.905357 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.922405 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.927573 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7495f97fdc-6pbrw"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937010 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-logs\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937062 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937101 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-combined-ca-bundle\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937122 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937149 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-config-data\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937187 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b719757-21ff-4967-a478-287c671d2bc7-logs\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937222 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-config-data\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937241 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937276 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fllcl\" (UniqueName: \"kubernetes.io/projected/5b719757-21ff-4967-a478-287c671d2bc7-kube-api-access-fllcl\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937299 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-scripts\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937337 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-combined-ca-bundle\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937370 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7stw\" (UniqueName: \"kubernetes.io/projected/d87862cf-9282-489b-b444-2b01b36e0999-kube-api-access-f7stw\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937390 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-db-sync-config-data\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937418 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937453 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z57n\" (UniqueName: \"kubernetes.io/projected/05565eb0-90db-4ec7-b646-db528fca40ad-kube-api-access-5z57n\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937491 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-scripts\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.937970 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b719757-21ff-4967-a478-287c671d2bc7-logs\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.959745 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.960980 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-combined-ca-bundle\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.980865 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-scripts\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.981246 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-db-sync-config-data\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.986732 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-596d946f85-qdrjh"] Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.987448 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-combined-ca-bundle\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:12 crc kubenswrapper[4492]: I0220 06:56:12.988363 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.005401 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z57n\" (UniqueName: \"kubernetes.io/projected/05565eb0-90db-4ec7-b646-db528fca40ad-kube-api-access-5z57n\") pod \"barbican-db-sync-d79gm\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.008280 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-config-data\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.009989 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fllcl\" (UniqueName: \"kubernetes.io/projected/5b719757-21ff-4967-a478-287c671d2bc7-kube-api-access-fllcl\") pod \"placement-db-sync-gjtqr\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.032010 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-596d946f85-qdrjh"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039675 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-logs\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039742 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039779 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-config\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039823 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmslp\" (UniqueName: \"kubernetes.io/projected/9a1f3e39-8b98-4bce-ae51-fa786e589c02-kube-api-access-vmslp\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039849 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039889 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-config-data\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039919 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-swift-storage-0\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.039975 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-nb\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.040018 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.040057 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-svc\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.040107 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-sb\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.040131 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7stw\" (UniqueName: \"kubernetes.io/projected/d87862cf-9282-489b-b444-2b01b36e0999-kube-api-access-f7stw\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.040170 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.040240 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-scripts\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.041333 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-logs\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.041576 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.045373 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.056959 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-scripts\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.066878 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7stw\" (UniqueName: \"kubernetes.io/projected/d87862cf-9282-489b-b444-2b01b36e0999-kube-api-access-f7stw\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.074243 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.083027 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.099527 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-config-data\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.123643 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.124463 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.144079 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7495f97fdc-6pbrw"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.146675 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-config\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.146791 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmslp\" (UniqueName: \"kubernetes.io/projected/9a1f3e39-8b98-4bce-ae51-fa786e589c02-kube-api-access-vmslp\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.148836 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-swift-storage-0\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.148942 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-nb\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.149044 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-svc\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.149129 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-sb\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.149904 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-sb\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.148944 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-config\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.150201 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-nb\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.150740 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-svc\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.151431 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-swift-storage-0\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.175194 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d79gm" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.182903 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmslp\" (UniqueName: \"kubernetes.io/projected/9a1f3e39-8b98-4bce-ae51-fa786e589c02-kube-api-access-vmslp\") pod \"dnsmasq-dns-596d946f85-qdrjh\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.186334 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gjtqr" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.209139 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.210036 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.216423 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:13 crc kubenswrapper[4492]: E0220 06:56:13.216905 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerName="init" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.216921 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerName="init" Feb 20 06:56:13 crc kubenswrapper[4492]: E0220 06:56:13.216938 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerName="dnsmasq-dns" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.216945 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerName="dnsmasq-dns" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.217129 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" containerName="dnsmasq-dns" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.218720 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.225041 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.225132 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.255069 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.342131 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.351927 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" event={"ID":"cca7a0af-f44b-43f8-8f45-894ee3609a0a","Type":"ContainerDied","Data":"b96bf847e3e2b503106bbe09e88d7307af6bbaa24054d01881748bce880070fc"} Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.351978 4492 scope.go:117] "RemoveContainer" containerID="f7ee6c3fee157ec60a72a08f540ad9f3e5137b84d114ba4079b7b7b27552668f" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.352042 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8668cb6c6f-2bknz" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.353307 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-config\") pod \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.353358 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkj9d\" (UniqueName: \"kubernetes.io/projected/cca7a0af-f44b-43f8-8f45-894ee3609a0a-kube-api-access-nkj9d\") pod \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.353433 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-sb\") pod \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.353678 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-svc\") pod \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.353717 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-swift-storage-0\") pod \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.353758 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-nb\") pod \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\" (UID: \"cca7a0af-f44b-43f8-8f45-894ee3609a0a\") " Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354064 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354139 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354158 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354182 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354228 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-logs\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354251 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354304 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.354324 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bwzf\" (UniqueName: \"kubernetes.io/projected/fafda38f-3eb5-4048-9f27-9b3a9028cc47-kube-api-access-7bwzf\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.395560 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z7d95" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="registry-server" containerID="cri-o://a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097" gracePeriod=2 Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.395822 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" event={"ID":"870fda83-5faf-49c8-8b92-4ac16a116677","Type":"ContainerStarted","Data":"96e1a2ef6d7da622485992ad97f19bcc58a7d69b25a97fceb46ced839586ddf3"} Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.413051 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca7a0af-f44b-43f8-8f45-894ee3609a0a-kube-api-access-nkj9d" (OuterVolumeSpecName: "kube-api-access-nkj9d") pod "cca7a0af-f44b-43f8-8f45-894ee3609a0a" (UID: "cca7a0af-f44b-43f8-8f45-894ee3609a0a"). InnerVolumeSpecName "kube-api-access-nkj9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.461502 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.461980 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.462007 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.462054 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.462161 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-logs\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.462241 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.462580 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.462673 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bwzf\" (UniqueName: \"kubernetes.io/projected/fafda38f-3eb5-4048-9f27-9b3a9028cc47-kube-api-access-7bwzf\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.463998 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-logs\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.464760 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.468698 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkj9d\" (UniqueName: \"kubernetes.io/projected/cca7a0af-f44b-43f8-8f45-894ee3609a0a-kube-api-access-nkj9d\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.471444 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.484082 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cca7a0af-f44b-43f8-8f45-894ee3609a0a" (UID: "cca7a0af-f44b-43f8-8f45-894ee3609a0a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.484096 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cca7a0af-f44b-43f8-8f45-894ee3609a0a" (UID: "cca7a0af-f44b-43f8-8f45-894ee3609a0a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.484858 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.506838 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.507635 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-config" (OuterVolumeSpecName: "config") pod "cca7a0af-f44b-43f8-8f45-894ee3609a0a" (UID: "cca7a0af-f44b-43f8-8f45-894ee3609a0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.526181 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.526194 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bwzf\" (UniqueName: \"kubernetes.io/projected/fafda38f-3eb5-4048-9f27-9b3a9028cc47-kube-api-access-7bwzf\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.531367 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.542637 4492 scope.go:117] "RemoveContainer" containerID="e9c3e15217b9ab793bffd909053306530a37c43aad80d26d3218df2ed4c849bf" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.572493 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.572524 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.572535 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.585206 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cca7a0af-f44b-43f8-8f45-894ee3609a0a" (UID: "cca7a0af-f44b-43f8-8f45-894ee3609a0a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.598599 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.637743 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cca7a0af-f44b-43f8-8f45-894ee3609a0a" (UID: "cca7a0af-f44b-43f8-8f45-894ee3609a0a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.680550 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.680787 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cca7a0af-f44b-43f8-8f45-894ee3609a0a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:13 crc kubenswrapper[4492]: W0220 06:56:13.759463 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aa62fa4_cbb7_4a89_8de1_91eac6703146.slice/crio-1e4c67ce34b94243487ac038194b0dcd7be47ed08e8e043802021ec50580d831 WatchSource:0}: Error finding container 1e4c67ce34b94243487ac038194b0dcd7be47ed08e8e043802021ec50580d831: Status 404 returned error can't find the container with id 1e4c67ce34b94243487ac038194b0dcd7be47ed08e8e043802021ec50580d831 Feb 20 06:56:13 crc kubenswrapper[4492]: W0220 06:56:13.775514 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41a3dc5d_8b2a_4be3_883c_d84d7a5c1859.slice/crio-a4064f9356808014ae3d3c8748ed1229ef302709efecf3fd07e79712133f28fa WatchSource:0}: Error finding container a4064f9356808014ae3d3c8748ed1229ef302709efecf3fd07e79712133f28fa: Status 404 returned error can't find the container with id a4064f9356808014ae3d3c8748ed1229ef302709efecf3fd07e79712133f28fa Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.848849 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-s246x"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.849083 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-h6p7g"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.852752 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-qqhvc"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.854372 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.911741 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6554ff645f-r4p2z"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.925299 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8668cb6c6f-2bknz"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.944435 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8668cb6c6f-2bknz"] Feb 20 06:56:13 crc kubenswrapper[4492]: I0220 06:56:13.994570 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d5c9766b9-5tjrb"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.017355 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2s268"] Feb 20 06:56:14 crc kubenswrapper[4492]: W0220 06:56:14.017522 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d697f3d_7b65_4983_8602_de839d5caa82.slice/crio-2499b1afd901cf603e5549289059a534e6cde4af4d321f3c247d53a4a91127f0 WatchSource:0}: Error finding container 2499b1afd901cf603e5549289059a534e6cde4af4d321f3c247d53a4a91127f0: Status 404 returned error can't find the container with id 2499b1afd901cf603e5549289059a534e6cde4af4d321f3c247d53a4a91127f0 Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.177625 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.218628 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-d79gm"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.364774 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.389521 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.458657 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h6p7g" event={"ID":"9aa62fa4-cbb7-4a89-8de1-91eac6703146","Type":"ContainerStarted","Data":"147dc48098e65dcc3fa6a521780d7102c598513b72fbd316186df5c77c6972b1"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.458711 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h6p7g" event={"ID":"9aa62fa4-cbb7-4a89-8de1-91eac6703146","Type":"ContainerStarted","Data":"1e4c67ce34b94243487ac038194b0dcd7be47ed08e8e043802021ec50580d831"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.476985 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6554ff645f-r4p2z"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.480637 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d5c9766b9-5tjrb" event={"ID":"b83b1978-9929-413b-a7b2-6f9bacf9af02","Type":"ContainerStarted","Data":"7bbe7d849b06e2986abf24d0c03fe8298f9338992a70c5ed3f6f70fc5506e931"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.512629 4492 generic.go:334] "Generic (PLEG): container finished" podID="870fda83-5faf-49c8-8b92-4ac16a116677" containerID="073f9fbce45b6f488d6222e68a00bf7105030e37f10ab247fb03e5bb801a75f3" exitCode=0 Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.512696 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" event={"ID":"870fda83-5faf-49c8-8b92-4ac16a116677","Type":"ContainerDied","Data":"073f9fbce45b6f488d6222e68a00bf7105030e37f10ab247fb03e5bb801a75f3"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.530809 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d79gm" event={"ID":"05565eb0-90db-4ec7-b646-db528fca40ad","Type":"ContainerStarted","Data":"ec2315be0308f36c13788159dace69c10acd2d1b387f2ee8678d5ba7269fe415"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.531698 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-utilities\") pod \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.531859 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28cvr\" (UniqueName: \"kubernetes.io/projected/6617a084-7a2b-4b2a-bec2-d1eb4378c841-kube-api-access-28cvr\") pod \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.532007 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-catalog-content\") pod \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\" (UID: \"6617a084-7a2b-4b2a-bec2-d1eb4378c841\") " Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.533501 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-h6p7g" podStartSLOduration=3.533454924 podStartE2EDuration="3.533454924s" podCreationTimestamp="2026-02-20 06:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:14.498268758 +0000 UTC m=+931.269557737" watchObservedRunningTime="2026-02-20 06:56:14.533454924 +0000 UTC m=+931.304743902" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.533983 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-utilities" (OuterVolumeSpecName: "utilities") pod "6617a084-7a2b-4b2a-bec2-d1eb4378c841" (UID: "6617a084-7a2b-4b2a-bec2-d1eb4378c841"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.534342 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7cdf7bb8f-dkm64"] Feb 20 06:56:14 crc kubenswrapper[4492]: E0220 06:56:14.534736 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="registry-server" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.534755 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="registry-server" Feb 20 06:56:14 crc kubenswrapper[4492]: E0220 06:56:14.534783 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="extract-content" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.534789 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="extract-content" Feb 20 06:56:14 crc kubenswrapper[4492]: E0220 06:56:14.534817 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="extract-utilities" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.534824 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="extract-utilities" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.534997 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerName="registry-server" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.535834 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.557098 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qqhvc" event={"ID":"82ed151d-241b-465d-8b5d-d97fe51015df","Type":"ContainerStarted","Data":"0ab7743ef909a89d2ed691be403a464f86185976b1dc84851ce3eddd12cfa308"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.579902 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.587319 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6617a084-7a2b-4b2a-bec2-d1eb4378c841-kube-api-access-28cvr" (OuterVolumeSpecName: "kube-api-access-28cvr") pod "6617a084-7a2b-4b2a-bec2-d1eb4378c841" (UID: "6617a084-7a2b-4b2a-bec2-d1eb4378c841"). InnerVolumeSpecName "kube-api-access-28cvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.606550 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cdf7bb8f-dkm64"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.615141 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-s246x" event={"ID":"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc","Type":"ContainerStarted","Data":"6121ddd77210f62d523ad6eb8cfb464a2641e64d6a21479ef7e06c553c32a9ca"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.638033 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm8tw\" (UniqueName: \"kubernetes.io/projected/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-kube-api-access-hm8tw\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.638075 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-logs\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.638160 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-config-data\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.638227 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-scripts\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.638305 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-horizon-secret-key\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.638428 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28cvr\" (UniqueName: \"kubernetes.io/projected/6617a084-7a2b-4b2a-bec2-d1eb4378c841-kube-api-access-28cvr\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.638441 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.652576 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6617a084-7a2b-4b2a-bec2-d1eb4378c841" (UID: "6617a084-7a2b-4b2a-bec2-d1eb4378c841"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.677106 4492 generic.go:334] "Generic (PLEG): container finished" podID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" containerID="a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097" exitCode=0 Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.677457 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z7d95" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.677772 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gjtqr"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.677828 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7d95" event={"ID":"6617a084-7a2b-4b2a-bec2-d1eb4378c841","Type":"ContainerDied","Data":"a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.677859 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z7d95" event={"ID":"6617a084-7a2b-4b2a-bec2-d1eb4378c841","Type":"ContainerDied","Data":"fa0d81279a8206274ca8aa4cc48eec6dda96e116396044da30993789d4e8ab90"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.677880 4492 scope.go:117] "RemoveContainer" containerID="a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.694213 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2s268" event={"ID":"4d697f3d-7b65-4983-8602-de839d5caa82","Type":"ContainerStarted","Data":"7783b346b02ae90a0e391645933bc3abef3edc4f1a4e94804ba624f030168e71"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.694251 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2s268" event={"ID":"4d697f3d-7b65-4983-8602-de839d5caa82","Type":"ContainerStarted","Data":"2499b1afd901cf603e5549289059a534e6cde4af4d321f3c247d53a4a91127f0"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.703753 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6554ff645f-r4p2z" event={"ID":"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859","Type":"ContainerStarted","Data":"a4064f9356808014ae3d3c8748ed1229ef302709efecf3fd07e79712133f28fa"} Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.722355 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerStarted","Data":"6623062694fe630efe7592edb65f3cd7073b983a25176c0eb4c520c7cf601154"} Feb 20 06:56:14 crc kubenswrapper[4492]: W0220 06:56:14.724300 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd87862cf_9282_489b_b444_2b01b36e0999.slice/crio-37c3f067003e82f42b022a43197849c36ca7fada23dfe30032103a4e62415dbe WatchSource:0}: Error finding container 37c3f067003e82f42b022a43197849c36ca7fada23dfe30032103a4e62415dbe: Status 404 returned error can't find the container with id 37c3f067003e82f42b022a43197849c36ca7fada23dfe30032103a4e62415dbe Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.734827 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.742636 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-horizon-secret-key\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.742817 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm8tw\" (UniqueName: \"kubernetes.io/projected/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-kube-api-access-hm8tw\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.742840 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-logs\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.742899 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-config-data\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.742956 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-scripts\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.743034 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6617a084-7a2b-4b2a-bec2-d1eb4378c841-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.743596 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-scripts\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.743842 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-logs\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.747438 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-config-data\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.782907 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.787697 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-horizon-secret-key\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.789025 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm8tw\" (UniqueName: \"kubernetes.io/projected/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-kube-api-access-hm8tw\") pod \"horizon-7cdf7bb8f-dkm64\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.793795 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2s268" podStartSLOduration=2.793776413 podStartE2EDuration="2.793776413s" podCreationTimestamp="2026-02-20 06:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:14.7380844 +0000 UTC m=+931.509373378" watchObservedRunningTime="2026-02-20 06:56:14.793776413 +0000 UTC m=+931.565065391" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.802714 4492 scope.go:117] "RemoveContainer" containerID="a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.876775 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7d95"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.903461 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.911130 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z7d95"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.954610 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-596d946f85-qdrjh"] Feb 20 06:56:14 crc kubenswrapper[4492]: I0220 06:56:14.964418 4492 scope.go:117] "RemoveContainer" containerID="0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.016224 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.209155 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.236936 4492 scope.go:117] "RemoveContainer" containerID="a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097" Feb 20 06:56:15 crc kubenswrapper[4492]: E0220 06:56:15.238602 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097\": container with ID starting with a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097 not found: ID does not exist" containerID="a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.238684 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097"} err="failed to get container status \"a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097\": rpc error: code = NotFound desc = could not find container \"a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097\": container with ID starting with a12fa317a198cfe4cafd7339cf1e4e1538250fef4a9059448f2acda0881dc097 not found: ID does not exist" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.238713 4492 scope.go:117] "RemoveContainer" containerID="a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9" Feb 20 06:56:15 crc kubenswrapper[4492]: E0220 06:56:15.248670 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9\": container with ID starting with a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9 not found: ID does not exist" containerID="a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.248709 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9"} err="failed to get container status \"a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9\": rpc error: code = NotFound desc = could not find container \"a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9\": container with ID starting with a4df20f04dff29c5b80e7d01afec92ca57b2a45f894e31c82f9dc79969776ce9 not found: ID does not exist" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.248737 4492 scope.go:117] "RemoveContainer" containerID="0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7" Feb 20 06:56:15 crc kubenswrapper[4492]: E0220 06:56:15.249217 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7\": container with ID starting with 0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7 not found: ID does not exist" containerID="0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.249241 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7"} err="failed to get container status \"0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7\": rpc error: code = NotFound desc = could not find container \"0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7\": container with ID starting with 0c9150b8ab718216a283dde0e5b073b430e5bedd811f092dcb7fe1f586e9b8d7 not found: ID does not exist" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.355170 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-swift-storage-0\") pod \"870fda83-5faf-49c8-8b92-4ac16a116677\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.355509 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-svc\") pod \"870fda83-5faf-49c8-8b92-4ac16a116677\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.355589 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-config\") pod \"870fda83-5faf-49c8-8b92-4ac16a116677\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.355668 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-nb\") pod \"870fda83-5faf-49c8-8b92-4ac16a116677\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.355696 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrclr\" (UniqueName: \"kubernetes.io/projected/870fda83-5faf-49c8-8b92-4ac16a116677-kube-api-access-rrclr\") pod \"870fda83-5faf-49c8-8b92-4ac16a116677\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.355720 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-sb\") pod \"870fda83-5faf-49c8-8b92-4ac16a116677\" (UID: \"870fda83-5faf-49c8-8b92-4ac16a116677\") " Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.404751 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/870fda83-5faf-49c8-8b92-4ac16a116677-kube-api-access-rrclr" (OuterVolumeSpecName: "kube-api-access-rrclr") pod "870fda83-5faf-49c8-8b92-4ac16a116677" (UID: "870fda83-5faf-49c8-8b92-4ac16a116677"). InnerVolumeSpecName "kube-api-access-rrclr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.426700 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "870fda83-5faf-49c8-8b92-4ac16a116677" (UID: "870fda83-5faf-49c8-8b92-4ac16a116677"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.474645 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-config" (OuterVolumeSpecName: "config") pod "870fda83-5faf-49c8-8b92-4ac16a116677" (UID: "870fda83-5faf-49c8-8b92-4ac16a116677"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.475326 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "870fda83-5faf-49c8-8b92-4ac16a116677" (UID: "870fda83-5faf-49c8-8b92-4ac16a116677"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.476996 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.477019 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.477030 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrclr\" (UniqueName: \"kubernetes.io/projected/870fda83-5faf-49c8-8b92-4ac16a116677-kube-api-access-rrclr\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.477044 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.478822 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "870fda83-5faf-49c8-8b92-4ac16a116677" (UID: "870fda83-5faf-49c8-8b92-4ac16a116677"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.503074 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "870fda83-5faf-49c8-8b92-4ac16a116677" (UID: "870fda83-5faf-49c8-8b92-4ac16a116677"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.592062 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.592104 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/870fda83-5faf-49c8-8b92-4ac16a116677-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.599796 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6617a084-7a2b-4b2a-bec2-d1eb4378c841" path="/var/lib/kubelet/pods/6617a084-7a2b-4b2a-bec2-d1eb4378c841/volumes" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.600697 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cca7a0af-f44b-43f8-8f45-894ee3609a0a" path="/var/lib/kubelet/pods/cca7a0af-f44b-43f8-8f45-894ee3609a0a/volumes" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.741764 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" event={"ID":"870fda83-5faf-49c8-8b92-4ac16a116677","Type":"ContainerDied","Data":"96e1a2ef6d7da622485992ad97f19bcc58a7d69b25a97fceb46ced839586ddf3"} Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.742123 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7495f97fdc-6pbrw" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.742166 4492 scope.go:117] "RemoveContainer" containerID="073f9fbce45b6f488d6222e68a00bf7105030e37f10ab247fb03e5bb801a75f3" Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.756157 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cdf7bb8f-dkm64"] Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.779519 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fafda38f-3eb5-4048-9f27-9b3a9028cc47","Type":"ContainerStarted","Data":"e9cff90ba0d6128b3d4e7c5802ed265608125f05791a8699f4f5030e6c115227"} Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.784693 4492 generic.go:334] "Generic (PLEG): container finished" podID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerID="e60b90c3bc6755170de6a3c42f38dfe317bf1783d765457ec558f41b2a5ddf34" exitCode=0 Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.784766 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" event={"ID":"9a1f3e39-8b98-4bce-ae51-fa786e589c02","Type":"ContainerDied","Data":"e60b90c3bc6755170de6a3c42f38dfe317bf1783d765457ec558f41b2a5ddf34"} Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.784797 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" event={"ID":"9a1f3e39-8b98-4bce-ae51-fa786e589c02","Type":"ContainerStarted","Data":"9675dc1eaf331252194c53f96bc0186656187594fd0fe60bafab9449d98f3c0a"} Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.800877 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gjtqr" event={"ID":"5b719757-21ff-4967-a478-287c671d2bc7","Type":"ContainerStarted","Data":"8ef0971b3fe24ea6d3b15ef40c8a9a0fd722bdd0b512fb31fe9f7352aedb7252"} Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.864944 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7495f97fdc-6pbrw"] Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.877367 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d87862cf-9282-489b-b444-2b01b36e0999","Type":"ContainerStarted","Data":"37c3f067003e82f42b022a43197849c36ca7fada23dfe30032103a4e62415dbe"} Feb 20 06:56:15 crc kubenswrapper[4492]: I0220 06:56:15.931762 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7495f97fdc-6pbrw"] Feb 20 06:56:16 crc kubenswrapper[4492]: I0220 06:56:16.928995 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" event={"ID":"9a1f3e39-8b98-4bce-ae51-fa786e589c02","Type":"ContainerStarted","Data":"35dbf0337debb0422acca4c4dbbd7060833037e4bbe0444696d53a3569b9eb2d"} Feb 20 06:56:16 crc kubenswrapper[4492]: I0220 06:56:16.929538 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:16 crc kubenswrapper[4492]: I0220 06:56:16.940501 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d87862cf-9282-489b-b444-2b01b36e0999","Type":"ContainerStarted","Data":"da1f81c37dc29e26c561d18931d35d6667a0d9062ffa4f5ae9bc3c86d436dd0c"} Feb 20 06:56:16 crc kubenswrapper[4492]: I0220 06:56:16.949794 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cdf7bb8f-dkm64" event={"ID":"bdcadbf5-6fc2-499b-acc8-f9aef43803dd","Type":"ContainerStarted","Data":"00650ee3a06d2789949e39287ab7d21560b3c4b95702c2d774347b4f946889a2"} Feb 20 06:56:16 crc kubenswrapper[4492]: I0220 06:56:16.957182 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" podStartSLOduration=4.957169157 podStartE2EDuration="4.957169157s" podCreationTimestamp="2026-02-20 06:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:16.954788691 +0000 UTC m=+933.726077669" watchObservedRunningTime="2026-02-20 06:56:16.957169157 +0000 UTC m=+933.728458135" Feb 20 06:56:17 crc kubenswrapper[4492]: I0220 06:56:17.570316 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="870fda83-5faf-49c8-8b92-4ac16a116677" path="/var/lib/kubelet/pods/870fda83-5faf-49c8-8b92-4ac16a116677/volumes" Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.064183 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fafda38f-3eb5-4048-9f27-9b3a9028cc47","Type":"ContainerStarted","Data":"fe2a41f9c005a16791c1c499565fd2d932074e2cead0b8f633ed6156c582b2eb"} Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.064285 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fafda38f-3eb5-4048-9f27-9b3a9028cc47","Type":"ContainerStarted","Data":"a1036b53158b7c365ebc24bb94bf86044f6e726b6d97668d10765ef055be08ac"} Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.064533 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-log" containerID="cri-o://fe2a41f9c005a16791c1c499565fd2d932074e2cead0b8f633ed6156c582b2eb" gracePeriod=30 Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.064741 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-httpd" containerID="cri-o://a1036b53158b7c365ebc24bb94bf86044f6e726b6d97668d10765ef055be08ac" gracePeriod=30 Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.120999 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.1209811179999996 podStartE2EDuration="6.120981118s" podCreationTimestamp="2026-02-20 06:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:18.118090692 +0000 UTC m=+934.889379670" watchObservedRunningTime="2026-02-20 06:56:18.120981118 +0000 UTC m=+934.892270096" Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.121846 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d87862cf-9282-489b-b444-2b01b36e0999","Type":"ContainerStarted","Data":"bcabc6f3f996af9ef5ee90c40113ce781e659a7a4c4b6a4e9ab2297a0db6c2a1"} Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.133665 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-log" containerID="cri-o://da1f81c37dc29e26c561d18931d35d6667a0d9062ffa4f5ae9bc3c86d436dd0c" gracePeriod=30 Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.134279 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-httpd" containerID="cri-o://bcabc6f3f996af9ef5ee90c40113ce781e659a7a4c4b6a4e9ab2297a0db6c2a1" gracePeriod=30 Feb 20 06:56:18 crc kubenswrapper[4492]: I0220 06:56:18.171278 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.1712605 podStartE2EDuration="6.1712605s" podCreationTimestamp="2026-02-20 06:56:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:18.1638105 +0000 UTC m=+934.935099478" watchObservedRunningTime="2026-02-20 06:56:18.1712605 +0000 UTC m=+934.942549479" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.153230 4492 generic.go:334] "Generic (PLEG): container finished" podID="d87862cf-9282-489b-b444-2b01b36e0999" containerID="bcabc6f3f996af9ef5ee90c40113ce781e659a7a4c4b6a4e9ab2297a0db6c2a1" exitCode=143 Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.153542 4492 generic.go:334] "Generic (PLEG): container finished" podID="d87862cf-9282-489b-b444-2b01b36e0999" containerID="da1f81c37dc29e26c561d18931d35d6667a0d9062ffa4f5ae9bc3c86d436dd0c" exitCode=143 Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.153394 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d87862cf-9282-489b-b444-2b01b36e0999","Type":"ContainerDied","Data":"bcabc6f3f996af9ef5ee90c40113ce781e659a7a4c4b6a4e9ab2297a0db6c2a1"} Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.153701 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d87862cf-9282-489b-b444-2b01b36e0999","Type":"ContainerDied","Data":"da1f81c37dc29e26c561d18931d35d6667a0d9062ffa4f5ae9bc3c86d436dd0c"} Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.153724 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d87862cf-9282-489b-b444-2b01b36e0999","Type":"ContainerDied","Data":"37c3f067003e82f42b022a43197849c36ca7fada23dfe30032103a4e62415dbe"} Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.153736 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37c3f067003e82f42b022a43197849c36ca7fada23dfe30032103a4e62415dbe" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.158651 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fafda38f-3eb5-4048-9f27-9b3a9028cc47","Type":"ContainerDied","Data":"a1036b53158b7c365ebc24bb94bf86044f6e726b6d97668d10765ef055be08ac"} Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.158603 4492 generic.go:334] "Generic (PLEG): container finished" podID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerID="a1036b53158b7c365ebc24bb94bf86044f6e726b6d97668d10765ef055be08ac" exitCode=143 Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.158705 4492 generic.go:334] "Generic (PLEG): container finished" podID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerID="fe2a41f9c005a16791c1c499565fd2d932074e2cead0b8f633ed6156c582b2eb" exitCode=143 Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.158720 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fafda38f-3eb5-4048-9f27-9b3a9028cc47","Type":"ContainerDied","Data":"fe2a41f9c005a16791c1c499565fd2d932074e2cead0b8f633ed6156c582b2eb"} Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.203146 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.329419 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-public-tls-certs\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.329723 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.329798 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-logs\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.329830 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-combined-ca-bundle\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.329858 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7stw\" (UniqueName: \"kubernetes.io/projected/d87862cf-9282-489b-b444-2b01b36e0999-kube-api-access-f7stw\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.329936 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-config-data\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.330026 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-httpd-run\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.330046 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-scripts\") pod \"d87862cf-9282-489b-b444-2b01b36e0999\" (UID: \"d87862cf-9282-489b-b444-2b01b36e0999\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.339823 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-logs" (OuterVolumeSpecName: "logs") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.339937 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.342568 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d87862cf-9282-489b-b444-2b01b36e0999-kube-api-access-f7stw" (OuterVolumeSpecName: "kube-api-access-f7stw") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "kube-api-access-f7stw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.349995 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.356297 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-scripts" (OuterVolumeSpecName: "scripts") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.376600 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.412690 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-config-data" (OuterVolumeSpecName: "config-data") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.418453 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d87862cf-9282-489b-b444-2b01b36e0999" (UID: "d87862cf-9282-489b-b444-2b01b36e0999"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440376 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440402 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440412 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440696 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440902 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440920 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d87862cf-9282-489b-b444-2b01b36e0999-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440934 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d87862cf-9282-489b-b444-2b01b36e0999-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.440944 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7stw\" (UniqueName: \"kubernetes.io/projected/d87862cf-9282-489b-b444-2b01b36e0999-kube-api-access-f7stw\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.512799 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.543188 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.860530 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959060 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-logs\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959110 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-internal-tls-certs\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959186 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959254 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-httpd-run\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959272 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-config-data\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959352 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-combined-ca-bundle\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959413 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-scripts\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.959441 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bwzf\" (UniqueName: \"kubernetes.io/projected/fafda38f-3eb5-4048-9f27-9b3a9028cc47-kube-api-access-7bwzf\") pod \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\" (UID: \"fafda38f-3eb5-4048-9f27-9b3a9028cc47\") " Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.960058 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.966204 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-logs" (OuterVolumeSpecName: "logs") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.966365 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fafda38f-3eb5-4048-9f27-9b3a9028cc47-kube-api-access-7bwzf" (OuterVolumeSpecName: "kube-api-access-7bwzf") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "kube-api-access-7bwzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.966586 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.968412 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-scripts" (OuterVolumeSpecName: "scripts") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:19 crc kubenswrapper[4492]: I0220 06:56:19.997992 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.012703 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.062876 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-config-data" (OuterVolumeSpecName: "config-data") pod "fafda38f-3eb5-4048-9f27-9b3a9028cc47" (UID: "fafda38f-3eb5-4048-9f27-9b3a9028cc47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.064741 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.065208 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.065292 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bwzf\" (UniqueName: \"kubernetes.io/projected/fafda38f-3eb5-4048-9f27-9b3a9028cc47-kube-api-access-7bwzf\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.065358 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.065409 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.065501 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.065555 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fafda38f-3eb5-4048-9f27-9b3a9028cc47-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.065666 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fafda38f-3eb5-4048-9f27-9b3a9028cc47-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.092850 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.167657 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.190169 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fafda38f-3eb5-4048-9f27-9b3a9028cc47","Type":"ContainerDied","Data":"e9cff90ba0d6128b3d4e7c5802ed265608125f05791a8699f4f5030e6c115227"} Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.190223 4492 scope.go:117] "RemoveContainer" containerID="a1036b53158b7c365ebc24bb94bf86044f6e726b6d97668d10765ef055be08ac" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.190353 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.204728 4492 generic.go:334] "Generic (PLEG): container finished" podID="9aa62fa4-cbb7-4a89-8de1-91eac6703146" containerID="147dc48098e65dcc3fa6a521780d7102c598513b72fbd316186df5c77c6972b1" exitCode=0 Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.204811 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.205161 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h6p7g" event={"ID":"9aa62fa4-cbb7-4a89-8de1-91eac6703146","Type":"ContainerDied","Data":"147dc48098e65dcc3fa6a521780d7102c598513b72fbd316186df5c77c6972b1"} Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.240844 4492 scope.go:117] "RemoveContainer" containerID="fe2a41f9c005a16791c1c499565fd2d932074e2cead0b8f633ed6156c582b2eb" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.293529 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.324316 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.336536 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351218 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351283 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: E0220 06:56:20.351679 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-log" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351699 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-log" Feb 20 06:56:20 crc kubenswrapper[4492]: E0220 06:56:20.351715 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-httpd" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351722 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-httpd" Feb 20 06:56:20 crc kubenswrapper[4492]: E0220 06:56:20.351733 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-httpd" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351740 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-httpd" Feb 20 06:56:20 crc kubenswrapper[4492]: E0220 06:56:20.351766 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="870fda83-5faf-49c8-8b92-4ac16a116677" containerName="init" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351772 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="870fda83-5faf-49c8-8b92-4ac16a116677" containerName="init" Feb 20 06:56:20 crc kubenswrapper[4492]: E0220 06:56:20.351780 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-log" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351786 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-log" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351944 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="870fda83-5faf-49c8-8b92-4ac16a116677" containerName="init" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351957 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-httpd" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351966 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-httpd" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351981 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" containerName="glance-log" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.351989 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87862cf-9282-489b-b444-2b01b36e0999" containerName="glance-log" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.361398 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.368981 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.383076 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jvgvq" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.408862 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.409332 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.409618 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.478989 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.479025 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.479135 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486455 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486556 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czf7l\" (UniqueName: \"kubernetes.io/projected/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-kube-api-access-czf7l\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486588 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486609 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486680 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486738 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-logs\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486792 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.486830 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.490147 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.490355 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.589162 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.589439 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.589678 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czf7l\" (UniqueName: \"kubernetes.io/projected/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-kube-api-access-czf7l\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.589772 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.589842 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.589926 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.590005 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ppk9\" (UniqueName: \"kubernetes.io/projected/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-kube-api-access-6ppk9\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.590080 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.590209 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.590334 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.590426 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-logs\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.590941 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.591045 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-logs\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.591233 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.591317 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.591384 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-logs\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.591406 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.591450 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.592106 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.604254 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.608121 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.609617 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.610297 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.618598 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czf7l\" (UniqueName: \"kubernetes.io/projected/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-kube-api-access-czf7l\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.637225 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.693897 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-logs\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.693939 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.694015 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.694036 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.694075 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.694098 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ppk9\" (UniqueName: \"kubernetes.io/projected/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-kube-api-access-6ppk9\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.694123 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.694273 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.696613 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.707779 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.708419 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-logs\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.708467 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.712517 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-scripts\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.721088 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.721420 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.721803 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-config-data\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.723717 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ppk9\" (UniqueName: \"kubernetes.io/projected/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-kube-api-access-6ppk9\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.769423 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.809005 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.868960 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d5c9766b9-5tjrb"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.912987 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-65f5db898d-bxntp"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.914383 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.921281 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65f5db898d-bxntp"] Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.926405 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 20 06:56:20 crc kubenswrapper[4492]: I0220 06:56:20.973281 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.005942 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-tls-certs\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.006020 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-config-data\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.006078 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zhb8\" (UniqueName: \"kubernetes.io/projected/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-kube-api-access-8zhb8\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.006107 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-secret-key\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.006165 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-combined-ca-bundle\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.006188 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-scripts\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.006209 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-logs\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.012417 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cdf7bb8f-dkm64"] Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.034799 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.041714 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d96664ccd-bltcr"] Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.045827 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.063186 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d96664ccd-bltcr"] Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108603 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-combined-ca-bundle\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108661 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-horizon-tls-certs\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108713 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zhb8\" (UniqueName: \"kubernetes.io/projected/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-kube-api-access-8zhb8\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108742 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-secret-key\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108787 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22416222-576a-4b0c-a659-145ec65738e3-logs\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108814 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-combined-ca-bundle\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108839 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-scripts\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108856 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-logs\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108922 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-horizon-secret-key\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108968 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxkd6\" (UniqueName: \"kubernetes.io/projected/22416222-576a-4b0c-a659-145ec65738e3-kube-api-access-vxkd6\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.108983 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22416222-576a-4b0c-a659-145ec65738e3-config-data\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.109008 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22416222-576a-4b0c-a659-145ec65738e3-scripts\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.109058 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-tls-certs\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.109113 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-config-data\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.110832 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-config-data\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.111277 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-scripts\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.111517 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-logs\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.115242 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-combined-ca-bundle\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.118691 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-secret-key\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.122944 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-tls-certs\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.126652 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zhb8\" (UniqueName: \"kubernetes.io/projected/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-kube-api-access-8zhb8\") pod \"horizon-65f5db898d-bxntp\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211170 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-horizon-secret-key\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211220 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxkd6\" (UniqueName: \"kubernetes.io/projected/22416222-576a-4b0c-a659-145ec65738e3-kube-api-access-vxkd6\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211242 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22416222-576a-4b0c-a659-145ec65738e3-config-data\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211264 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22416222-576a-4b0c-a659-145ec65738e3-scripts\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211326 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-combined-ca-bundle\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211344 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-horizon-tls-certs\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211397 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22416222-576a-4b0c-a659-145ec65738e3-logs\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.211752 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22416222-576a-4b0c-a659-145ec65738e3-logs\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.213236 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22416222-576a-4b0c-a659-145ec65738e3-scripts\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.216891 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-horizon-secret-key\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.217278 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/22416222-576a-4b0c-a659-145ec65738e3-config-data\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.224729 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-combined-ca-bundle\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.225150 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/22416222-576a-4b0c-a659-145ec65738e3-horizon-tls-certs\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.227327 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxkd6\" (UniqueName: \"kubernetes.io/projected/22416222-576a-4b0c-a659-145ec65738e3-kube-api-access-vxkd6\") pod \"horizon-7d96664ccd-bltcr\" (UID: \"22416222-576a-4b0c-a659-145ec65738e3\") " pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.265556 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.418534 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.566579 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d87862cf-9282-489b-b444-2b01b36e0999" path="/var/lib/kubelet/pods/d87862cf-9282-489b-b444-2b01b36e0999/volumes" Feb 20 06:56:21 crc kubenswrapper[4492]: I0220 06:56:21.567538 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fafda38f-3eb5-4048-9f27-9b3a9028cc47" path="/var/lib/kubelet/pods/fafda38f-3eb5-4048-9f27-9b3a9028cc47/volumes" Feb 20 06:56:23 crc kubenswrapper[4492]: I0220 06:56:23.344978 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:56:23 crc kubenswrapper[4492]: I0220 06:56:23.475773 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cd7cb8bfc-82g6n"] Feb 20 06:56:23 crc kubenswrapper[4492]: I0220 06:56:23.476129 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="dnsmasq-dns" containerID="cri-o://34b2244167dc065019507b9fa083547fab46bc50dc3f9117bf27d380d6bdb160" gracePeriod=10 Feb 20 06:56:24 crc kubenswrapper[4492]: I0220 06:56:24.286428 4492 generic.go:334] "Generic (PLEG): container finished" podID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerID="34b2244167dc065019507b9fa083547fab46bc50dc3f9117bf27d380d6bdb160" exitCode=0 Feb 20 06:56:24 crc kubenswrapper[4492]: I0220 06:56:24.286704 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" event={"ID":"c2ecdc8e-1e7a-4b25-9989-44978951e22f","Type":"ContainerDied","Data":"34b2244167dc065019507b9fa083547fab46bc50dc3f9117bf27d380d6bdb160"} Feb 20 06:56:27 crc kubenswrapper[4492]: I0220 06:56:27.324452 4492 generic.go:334] "Generic (PLEG): container finished" podID="4d697f3d-7b65-4983-8602-de839d5caa82" containerID="7783b346b02ae90a0e391645933bc3abef3edc4f1a4e94804ba624f030168e71" exitCode=0 Feb 20 06:56:27 crc kubenswrapper[4492]: I0220 06:56:27.324518 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2s268" event={"ID":"4d697f3d-7b65-4983-8602-de839d5caa82","Type":"ContainerDied","Data":"7783b346b02ae90a0e391645933bc3abef3edc4f1a4e94804ba624f030168e71"} Feb 20 06:56:31 crc kubenswrapper[4492]: I0220 06:56:31.937418 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: i/o timeout" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.863811 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.864329 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.864504 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n545hcdh68dh75h58h655h6h559h585h667h5f4h647h87hc4h547hd8h5bfh65ch56dh54h5c8h5h565h68dh67ch647h596hddh5f4h557h5d6hb5q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dr62s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6554ff645f-r4p2z_openstack(41a3dc5d-8b2a-4be3-883c-d84d7a5c1859): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.870955 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923\\\"\"]" pod="openstack/horizon-6554ff645f-r4p2z" podUID="41a3dc5d-8b2a-4be3-883c-d84d7a5c1859" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.878277 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.878302 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.878375 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n687h65fh5c8h685h697h66fh54h5d7h96h67h686h5fbh5ffhb9h5bh7bh57dhdh567h54h65bh74h544h99h94h6fh668h599hddhddh65bh7bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g6v9s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5d5c9766b9-5tjrb_openstack(b83b1978-9929-413b-a7b2-6f9bacf9af02): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:56:36 crc kubenswrapper[4492]: E0220 06:56:36.880174 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923\\\"\"]" pod="openstack/horizon-5d5c9766b9-5tjrb" podUID="b83b1978-9929-413b-a7b2-6f9bacf9af02" Feb 20 06:56:36 crc kubenswrapper[4492]: I0220 06:56:36.938424 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: i/o timeout" Feb 20 06:56:36 crc kubenswrapper[4492]: I0220 06:56:36.939684 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.051310 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-config-data\") pod \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.051541 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-scripts\") pod \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.051744 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwvm6\" (UniqueName: \"kubernetes.io/projected/9aa62fa4-cbb7-4a89-8de1-91eac6703146-kube-api-access-dwvm6\") pod \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.051829 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-combined-ca-bundle\") pod \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.051901 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-credential-keys\") pod \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.051930 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-fernet-keys\") pod \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\" (UID: \"9aa62fa4-cbb7-4a89-8de1-91eac6703146\") " Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.057687 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa62fa4-cbb7-4a89-8de1-91eac6703146-kube-api-access-dwvm6" (OuterVolumeSpecName: "kube-api-access-dwvm6") pod "9aa62fa4-cbb7-4a89-8de1-91eac6703146" (UID: "9aa62fa4-cbb7-4a89-8de1-91eac6703146"). InnerVolumeSpecName "kube-api-access-dwvm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.058028 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-scripts" (OuterVolumeSpecName: "scripts") pod "9aa62fa4-cbb7-4a89-8de1-91eac6703146" (UID: "9aa62fa4-cbb7-4a89-8de1-91eac6703146"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.058707 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9aa62fa4-cbb7-4a89-8de1-91eac6703146" (UID: "9aa62fa4-cbb7-4a89-8de1-91eac6703146"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.063909 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9aa62fa4-cbb7-4a89-8de1-91eac6703146" (UID: "9aa62fa4-cbb7-4a89-8de1-91eac6703146"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.076178 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-config-data" (OuterVolumeSpecName: "config-data") pod "9aa62fa4-cbb7-4a89-8de1-91eac6703146" (UID: "9aa62fa4-cbb7-4a89-8de1-91eac6703146"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.081637 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9aa62fa4-cbb7-4a89-8de1-91eac6703146" (UID: "9aa62fa4-cbb7-4a89-8de1-91eac6703146"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.114871 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7pcm9"] Feb 20 06:56:37 crc kubenswrapper[4492]: E0220 06:56:37.115309 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa62fa4-cbb7-4a89-8de1-91eac6703146" containerName="keystone-bootstrap" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.115323 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa62fa4-cbb7-4a89-8de1-91eac6703146" containerName="keystone-bootstrap" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.115560 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa62fa4-cbb7-4a89-8de1-91eac6703146" containerName="keystone-bootstrap" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.121443 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.164692 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwvm6\" (UniqueName: \"kubernetes.io/projected/9aa62fa4-cbb7-4a89-8de1-91eac6703146-kube-api-access-dwvm6\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.164803 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.164877 4492 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.164938 4492 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.164998 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.165056 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9aa62fa4-cbb7-4a89-8de1-91eac6703146-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.171929 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7pcm9"] Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.265753 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-utilities\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.265874 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-catalog-content\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.265913 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7dcr\" (UniqueName: \"kubernetes.io/projected/3ed937a1-e598-4418-81f6-cbbd240e4d4a-kube-api-access-b7dcr\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.367400 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7dcr\" (UniqueName: \"kubernetes.io/projected/3ed937a1-e598-4418-81f6-cbbd240e4d4a-kube-api-access-b7dcr\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.367467 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-utilities\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.367599 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-catalog-content\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.368032 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-catalog-content\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.368108 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-utilities\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.383931 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7dcr\" (UniqueName: \"kubernetes.io/projected/3ed937a1-e598-4418-81f6-cbbd240e4d4a-kube-api-access-b7dcr\") pod \"redhat-operators-7pcm9\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.422320 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-h6p7g" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.423461 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-h6p7g" event={"ID":"9aa62fa4-cbb7-4a89-8de1-91eac6703146","Type":"ContainerDied","Data":"1e4c67ce34b94243487ac038194b0dcd7be47ed08e8e043802021ec50580d831"} Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.423545 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e4c67ce34b94243487ac038194b0dcd7be47ed08e8e043802021ec50580d831" Feb 20 06:56:37 crc kubenswrapper[4492]: I0220 06:56:37.456003 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.028121 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-h6p7g"] Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.040374 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-h6p7g"] Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.131455 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-blbr6"] Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.132566 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.136255 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4v2r8" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.136648 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.136827 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.136962 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.137083 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.154656 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-blbr6"] Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.280465 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-combined-ca-bundle\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.280597 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-credential-keys\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.280653 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b2qh\" (UniqueName: \"kubernetes.io/projected/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-kube-api-access-4b2qh\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.280707 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-scripts\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.280812 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-config-data\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.280867 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-fernet-keys\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.383030 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-config-data\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.383095 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-fernet-keys\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.383197 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-combined-ca-bundle\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.383252 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-credential-keys\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.383274 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b2qh\" (UniqueName: \"kubernetes.io/projected/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-kube-api-access-4b2qh\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.383294 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-scripts\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.388931 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-scripts\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.389185 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-combined-ca-bundle\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.390162 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-config-data\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.403856 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-credential-keys\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.404287 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-fernet-keys\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.404263 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b2qh\" (UniqueName: \"kubernetes.io/projected/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-kube-api-access-4b2qh\") pod \"keystone-bootstrap-blbr6\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:38 crc kubenswrapper[4492]: I0220 06:56:38.460830 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.315581 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.315658 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.315708 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.316876 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"721ada477bd6e659aca9b99823e1dcf15d8562664d9f214bcd56e2ac3871cc6b"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.316937 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://721ada477bd6e659aca9b99823e1dcf15d8562664d9f214bcd56e2ac3871cc6b" gracePeriod=600 Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.455310 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="721ada477bd6e659aca9b99823e1dcf15d8562664d9f214bcd56e2ac3871cc6b" exitCode=0 Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.455358 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"721ada477bd6e659aca9b99823e1dcf15d8562664d9f214bcd56e2ac3871cc6b"} Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.455395 4492 scope.go:117] "RemoveContainer" containerID="5090e1416f91468b7a37c5f3132c069b8613e7ce0e4c8c57bc77b7ef73cc9bce" Feb 20 06:56:39 crc kubenswrapper[4492]: I0220 06:56:39.565582 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa62fa4-cbb7-4a89-8de1-91eac6703146" path="/var/lib/kubelet/pods/9aa62fa4-cbb7-4a89-8de1-91eac6703146/volumes" Feb 20 06:56:39 crc kubenswrapper[4492]: E0220 06:56:39.725269 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:39 crc kubenswrapper[4492]: E0220 06:56:39.725745 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:39 crc kubenswrapper[4492]: E0220 06:56:39.725927 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n74hch679hc8hb6h5c9hc9h696h5b7h59fh5c9h8hd9h67fh667hfdh9h69h65chb7h69h575h98hc5h578h6h5d8h65dhc9h659h668h56cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hm8tw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7cdf7bb8f-dkm64_openstack(bdcadbf5-6fc2-499b-acc8-f9aef43803dd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:56:39 crc kubenswrapper[4492]: E0220 06:56:39.729137 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:f0473f3e18dd17d7021c02e991298923\\\"\"]" pod="openstack/horizon-7cdf7bb8f-dkm64" podUID="bdcadbf5-6fc2-499b-acc8-f9aef43803dd" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.357420 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.360386 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481238 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c6fj\" (UniqueName: \"kubernetes.io/projected/4d697f3d-7b65-4983-8602-de839d5caa82-kube-api-access-2c6fj\") pod \"4d697f3d-7b65-4983-8602-de839d5caa82\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481362 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8qtq\" (UniqueName: \"kubernetes.io/projected/c2ecdc8e-1e7a-4b25-9989-44978951e22f-kube-api-access-h8qtq\") pod \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481445 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-combined-ca-bundle\") pod \"4d697f3d-7b65-4983-8602-de839d5caa82\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481588 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-sb\") pod \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481690 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-nb\") pod \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481719 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-svc\") pod \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481743 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-config\") pod \"4d697f3d-7b65-4983-8602-de839d5caa82\" (UID: \"4d697f3d-7b65-4983-8602-de839d5caa82\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481789 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-swift-storage-0\") pod \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.481835 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-config\") pod \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\" (UID: \"c2ecdc8e-1e7a-4b25-9989-44978951e22f\") " Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.500400 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ecdc8e-1e7a-4b25-9989-44978951e22f-kube-api-access-h8qtq" (OuterVolumeSpecName: "kube-api-access-h8qtq") pod "c2ecdc8e-1e7a-4b25-9989-44978951e22f" (UID: "c2ecdc8e-1e7a-4b25-9989-44978951e22f"). InnerVolumeSpecName "kube-api-access-h8qtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.504032 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d697f3d-7b65-4983-8602-de839d5caa82-kube-api-access-2c6fj" (OuterVolumeSpecName: "kube-api-access-2c6fj") pod "4d697f3d-7b65-4983-8602-de839d5caa82" (UID: "4d697f3d-7b65-4983-8602-de839d5caa82"). InnerVolumeSpecName "kube-api-access-2c6fj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.505113 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" event={"ID":"c2ecdc8e-1e7a-4b25-9989-44978951e22f","Type":"ContainerDied","Data":"f9cb80cf7e788372fc64c0a0a481b5f4867bf0074f1fe7ba43ce12d4a8668bab"} Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.505394 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.510676 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2s268" event={"ID":"4d697f3d-7b65-4983-8602-de839d5caa82","Type":"ContainerDied","Data":"2499b1afd901cf603e5549289059a534e6cde4af4d321f3c247d53a4a91127f0"} Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.510734 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2499b1afd901cf603e5549289059a534e6cde4af4d321f3c247d53a4a91127f0" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.510794 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2s268" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.532005 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-config" (OuterVolumeSpecName: "config") pod "4d697f3d-7b65-4983-8602-de839d5caa82" (UID: "4d697f3d-7b65-4983-8602-de839d5caa82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.537622 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-config" (OuterVolumeSpecName: "config") pod "c2ecdc8e-1e7a-4b25-9989-44978951e22f" (UID: "c2ecdc8e-1e7a-4b25-9989-44978951e22f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.540023 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d697f3d-7b65-4983-8602-de839d5caa82" (UID: "4d697f3d-7b65-4983-8602-de839d5caa82"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.546390 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c2ecdc8e-1e7a-4b25-9989-44978951e22f" (UID: "c2ecdc8e-1e7a-4b25-9989-44978951e22f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.551143 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c2ecdc8e-1e7a-4b25-9989-44978951e22f" (UID: "c2ecdc8e-1e7a-4b25-9989-44978951e22f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.560204 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c2ecdc8e-1e7a-4b25-9989-44978951e22f" (UID: "c2ecdc8e-1e7a-4b25-9989-44978951e22f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.561569 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c2ecdc8e-1e7a-4b25-9989-44978951e22f" (UID: "c2ecdc8e-1e7a-4b25-9989-44978951e22f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583777 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c6fj\" (UniqueName: \"kubernetes.io/projected/4d697f3d-7b65-4983-8602-de839d5caa82-kube-api-access-2c6fj\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583815 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8qtq\" (UniqueName: \"kubernetes.io/projected/c2ecdc8e-1e7a-4b25-9989-44978951e22f-kube-api-access-h8qtq\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583906 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583917 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583927 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583938 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583951 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4d697f3d-7b65-4983-8602-de839d5caa82-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583960 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.583969 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2ecdc8e-1e7a-4b25-9989-44978951e22f-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.837913 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cd7cb8bfc-82g6n"] Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.845506 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cd7cb8bfc-82g6n"] Feb 20 06:56:41 crc kubenswrapper[4492]: I0220 06:56:41.939593 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6cd7cb8bfc-82g6n" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: i/o timeout" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.734465 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75dfbfdb9-5flr6"] Feb 20 06:56:42 crc kubenswrapper[4492]: E0220 06:56:42.740158 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="init" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.740174 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="init" Feb 20 06:56:42 crc kubenswrapper[4492]: E0220 06:56:42.740184 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="dnsmasq-dns" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.740190 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="dnsmasq-dns" Feb 20 06:56:42 crc kubenswrapper[4492]: E0220 06:56:42.740201 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d697f3d-7b65-4983-8602-de839d5caa82" containerName="neutron-db-sync" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.740207 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d697f3d-7b65-4983-8602-de839d5caa82" containerName="neutron-db-sync" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.740400 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" containerName="dnsmasq-dns" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.740414 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d697f3d-7b65-4983-8602-de839d5caa82" containerName="neutron-db-sync" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.741374 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.783678 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75dfbfdb9-5flr6"] Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.814046 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-svc\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.814092 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-sb\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.814141 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-nb\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.814204 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-swift-storage-0\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.814260 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdlvm\" (UniqueName: \"kubernetes.io/projected/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-kube-api-access-rdlvm\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.814352 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-config\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.915879 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-svc\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.915929 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-sb\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.915974 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-nb\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.916041 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-swift-storage-0\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.916097 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdlvm\" (UniqueName: \"kubernetes.io/projected/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-kube-api-access-rdlvm\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.916191 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-config\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.916837 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-svc\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.916977 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-nb\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.917112 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-config\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.917579 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-swift-storage-0\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.923221 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-sb\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.952221 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdlvm\" (UniqueName: \"kubernetes.io/projected/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-kube-api-access-rdlvm\") pod \"dnsmasq-dns-75dfbfdb9-5flr6\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.973581 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cf7fbb848-5c8g6"] Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.975116 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.976701 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.977750 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ndkd9" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.991711 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.992010 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 20 06:56:42 crc kubenswrapper[4492]: I0220 06:56:42.994271 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cf7fbb848-5c8g6"] Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.021863 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-httpd-config\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.022013 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-ovndb-tls-certs\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.022097 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-combined-ca-bundle\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.022177 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-config\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.023433 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26nst\" (UniqueName: \"kubernetes.io/projected/3a568bef-11fc-479e-bf90-4c2030405335-kube-api-access-26nst\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.061665 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.125447 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26nst\" (UniqueName: \"kubernetes.io/projected/3a568bef-11fc-479e-bf90-4c2030405335-kube-api-access-26nst\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.125671 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-httpd-config\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.125726 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-ovndb-tls-certs\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.125757 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-combined-ca-bundle\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.125792 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-config\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.130401 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-config\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.131391 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-ovndb-tls-certs\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.133814 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-combined-ca-bundle\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.139690 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26nst\" (UniqueName: \"kubernetes.io/projected/3a568bef-11fc-479e-bf90-4c2030405335-kube-api-access-26nst\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.146029 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-httpd-config\") pod \"neutron-6cf7fbb848-5c8g6\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.313639 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:43 crc kubenswrapper[4492]: I0220 06:56:43.600973 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2ecdc8e-1e7a-4b25-9989-44978951e22f" path="/var/lib/kubelet/pods/c2ecdc8e-1e7a-4b25-9989-44978951e22f/volumes" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.753848 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66cc4998b5-ts9pl"] Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.755804 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.759389 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.760793 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.769655 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66cc4998b5-ts9pl"] Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.785087 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-config\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.785358 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-httpd-config\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.785397 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-internal-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.785417 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-ovndb-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.785483 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnl9q\" (UniqueName: \"kubernetes.io/projected/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-kube-api-access-fnl9q\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.785587 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-public-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.785622 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-combined-ca-bundle\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.887365 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-httpd-config\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.887419 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-internal-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.887442 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-ovndb-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.887508 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnl9q\" (UniqueName: \"kubernetes.io/projected/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-kube-api-access-fnl9q\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.887673 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-public-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.887713 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-combined-ca-bundle\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.887797 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-config\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.892767 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-config\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.894148 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-internal-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.894398 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-ovndb-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.896749 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-public-tls-certs\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.901453 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-httpd-config\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.904982 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnl9q\" (UniqueName: \"kubernetes.io/projected/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-kube-api-access-fnl9q\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:44 crc kubenswrapper[4492]: I0220 06:56:44.925144 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-combined-ca-bundle\") pod \"neutron-66cc4998b5-ts9pl\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:45 crc kubenswrapper[4492]: I0220 06:56:45.094154 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.031913 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.037949 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.132542 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-scripts\") pod \"b83b1978-9929-413b-a7b2-6f9bacf9af02\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.132916 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b83b1978-9929-413b-a7b2-6f9bacf9af02-horizon-secret-key\") pod \"b83b1978-9929-413b-a7b2-6f9bacf9af02\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133009 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-horizon-secret-key\") pod \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133075 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr62s\" (UniqueName: \"kubernetes.io/projected/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-kube-api-access-dr62s\") pod \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133138 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-logs\") pod \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133169 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-config-data\") pod \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133238 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-scripts" (OuterVolumeSpecName: "scripts") pod "b83b1978-9929-413b-a7b2-6f9bacf9af02" (UID: "b83b1978-9929-413b-a7b2-6f9bacf9af02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133627 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-logs" (OuterVolumeSpecName: "logs") pod "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859" (UID: "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133689 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-config-data\") pod \"b83b1978-9929-413b-a7b2-6f9bacf9af02\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133768 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6v9s\" (UniqueName: \"kubernetes.io/projected/b83b1978-9929-413b-a7b2-6f9bacf9af02-kube-api-access-g6v9s\") pod \"b83b1978-9929-413b-a7b2-6f9bacf9af02\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133819 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83b1978-9929-413b-a7b2-6f9bacf9af02-logs\") pod \"b83b1978-9929-413b-a7b2-6f9bacf9af02\" (UID: \"b83b1978-9929-413b-a7b2-6f9bacf9af02\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.133849 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-scripts\") pod \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\" (UID: \"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.134327 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-config-data" (OuterVolumeSpecName: "config-data") pod "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859" (UID: "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.134430 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-config-data" (OuterVolumeSpecName: "config-data") pod "b83b1978-9929-413b-a7b2-6f9bacf9af02" (UID: "b83b1978-9929-413b-a7b2-6f9bacf9af02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.134792 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b83b1978-9929-413b-a7b2-6f9bacf9af02-logs" (OuterVolumeSpecName: "logs") pod "b83b1978-9929-413b-a7b2-6f9bacf9af02" (UID: "b83b1978-9929-413b-a7b2-6f9bacf9af02"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.134909 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-scripts" (OuterVolumeSpecName: "scripts") pod "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859" (UID: "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.135310 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.135337 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.135352 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.135363 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b83b1978-9929-413b-a7b2-6f9bacf9af02-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.135371 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.135381 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b83b1978-9929-413b-a7b2-6f9bacf9af02-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.140147 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b83b1978-9929-413b-a7b2-6f9bacf9af02-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b83b1978-9929-413b-a7b2-6f9bacf9af02" (UID: "b83b1978-9929-413b-a7b2-6f9bacf9af02"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.140287 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b83b1978-9929-413b-a7b2-6f9bacf9af02-kube-api-access-g6v9s" (OuterVolumeSpecName: "kube-api-access-g6v9s") pod "b83b1978-9929-413b-a7b2-6f9bacf9af02" (UID: "b83b1978-9929-413b-a7b2-6f9bacf9af02"). InnerVolumeSpecName "kube-api-access-g6v9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.142634 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-kube-api-access-dr62s" (OuterVolumeSpecName: "kube-api-access-dr62s") pod "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859" (UID: "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859"). InnerVolumeSpecName "kube-api-access-dr62s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.157138 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859" (UID: "41a3dc5d-8b2a-4be3-883c-d84d7a5c1859"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.237435 4492 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b83b1978-9929-413b-a7b2-6f9bacf9af02-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.237462 4492 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.237660 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr62s\" (UniqueName: \"kubernetes.io/projected/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859-kube-api-access-dr62s\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.237683 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6v9s\" (UniqueName: \"kubernetes.io/projected/b83b1978-9929-413b-a7b2-6f9bacf9af02-kube-api-access-g6v9s\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.467939 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.467994 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.468120 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5z57n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-d79gm_openstack(05565eb0-90db-4ec7-b646-db528fca40ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.469613 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-d79gm" podUID="05565eb0-90db-4ec7-b646-db528fca40ad" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.501782 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.542407 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm8tw\" (UniqueName: \"kubernetes.io/projected/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-kube-api-access-hm8tw\") pod \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.542650 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-scripts\") pod \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.542822 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-config-data\") pod \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.542882 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-logs\") pod \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.543144 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-scripts" (OuterVolumeSpecName: "scripts") pod "bdcadbf5-6fc2-499b-acc8-f9aef43803dd" (UID: "bdcadbf5-6fc2-499b-acc8-f9aef43803dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.543165 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-horizon-secret-key\") pod \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\" (UID: \"bdcadbf5-6fc2-499b-acc8-f9aef43803dd\") " Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.543334 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-logs" (OuterVolumeSpecName: "logs") pod "bdcadbf5-6fc2-499b-acc8-f9aef43803dd" (UID: "bdcadbf5-6fc2-499b-acc8-f9aef43803dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.543797 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.543821 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.543993 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-config-data" (OuterVolumeSpecName: "config-data") pod "bdcadbf5-6fc2-499b-acc8-f9aef43803dd" (UID: "bdcadbf5-6fc2-499b-acc8-f9aef43803dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.547059 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-kube-api-access-hm8tw" (OuterVolumeSpecName: "kube-api-access-hm8tw") pod "bdcadbf5-6fc2-499b-acc8-f9aef43803dd" (UID: "bdcadbf5-6fc2-499b-acc8-f9aef43803dd"). InnerVolumeSpecName "kube-api-access-hm8tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.548326 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bdcadbf5-6fc2-499b-acc8-f9aef43803dd" (UID: "bdcadbf5-6fc2-499b-acc8-f9aef43803dd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.626038 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6554ff645f-r4p2z" event={"ID":"41a3dc5d-8b2a-4be3-883c-d84d7a5c1859","Type":"ContainerDied","Data":"a4064f9356808014ae3d3c8748ed1229ef302709efecf3fd07e79712133f28fa"} Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.626056 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6554ff645f-r4p2z" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.627214 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d5c9766b9-5tjrb" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.628058 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d5c9766b9-5tjrb" event={"ID":"b83b1978-9929-413b-a7b2-6f9bacf9af02","Type":"ContainerDied","Data":"7bbe7d849b06e2986abf24d0c03fe8298f9338992a70c5ed3f6f70fc5506e931"} Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.630176 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cdf7bb8f-dkm64" event={"ID":"bdcadbf5-6fc2-499b-acc8-f9aef43803dd","Type":"ContainerDied","Data":"00650ee3a06d2789949e39287ab7d21560b3c4b95702c2d774347b4f946889a2"} Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.630207 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cdf7bb8f-dkm64" Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.631625 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:f0473f3e18dd17d7021c02e991298923\\\"\"" pod="openstack/barbican-db-sync-d79gm" podUID="05565eb0-90db-4ec7-b646-db528fca40ad" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.649351 4492 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.649379 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hm8tw\" (UniqueName: \"kubernetes.io/projected/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-kube-api-access-hm8tw\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.649393 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bdcadbf5-6fc2-499b-acc8-f9aef43803dd-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.685263 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6554ff645f-r4p2z"] Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.697201 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6554ff645f-r4p2z"] Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.732534 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cdf7bb8f-dkm64"] Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.736551 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7cdf7bb8f-dkm64"] Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.747422 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5d5c9766b9-5tjrb"] Feb 20 06:56:51 crc kubenswrapper[4492]: I0220 06:56:51.751969 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5d5c9766b9-5tjrb"] Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.896398 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-ceilometer-central:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.896441 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-ceilometer-central:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:51 crc kubenswrapper[4492]: E0220 06:56:51.896591 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-ceilometer-central:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n55fh666h58ch688h644h5f5h86h98h5dh5dchffh554h85h5f9hb4h579h87h5b6hd5h6h685h574h567hb4h677h5b9h78h95h5f8h5d5h5fdh9dq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gb8wp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:56:53 crc kubenswrapper[4492]: E0220 06:56:53.031071 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:53 crc kubenswrapper[4492]: E0220 06:56:53.031416 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:f0473f3e18dd17d7021c02e991298923" Feb 20 06:56:53 crc kubenswrapper[4492]: E0220 06:56:53.031666 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:f0473f3e18dd17d7021c02e991298923,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vxg4m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-qqhvc_openstack(82ed151d-241b-465d-8b5d-d97fe51015df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 06:56:53 crc kubenswrapper[4492]: E0220 06:56:53.033113 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-qqhvc" podUID="82ed151d-241b-465d-8b5d-d97fe51015df" Feb 20 06:56:53 crc kubenswrapper[4492]: I0220 06:56:53.098394 4492 scope.go:117] "RemoveContainer" containerID="34b2244167dc065019507b9fa083547fab46bc50dc3f9117bf27d380d6bdb160" Feb 20 06:56:53 crc kubenswrapper[4492]: I0220 06:56:53.234619 4492 scope.go:117] "RemoveContainer" containerID="412a0e64dc08264ca9bd0d302ca74fe41e028d5c9c7ba8cca05550330285047f" Feb 20 06:56:53 crc kubenswrapper[4492]: I0220 06:56:53.583167 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41a3dc5d-8b2a-4be3-883c-d84d7a5c1859" path="/var/lib/kubelet/pods/41a3dc5d-8b2a-4be3-883c-d84d7a5c1859/volumes" Feb 20 06:56:53 crc kubenswrapper[4492]: I0220 06:56:53.584033 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b83b1978-9929-413b-a7b2-6f9bacf9af02" path="/var/lib/kubelet/pods/b83b1978-9929-413b-a7b2-6f9bacf9af02/volumes" Feb 20 06:56:53 crc kubenswrapper[4492]: I0220 06:56:53.590672 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdcadbf5-6fc2-499b-acc8-f9aef43803dd" path="/var/lib/kubelet/pods/bdcadbf5-6fc2-499b-acc8-f9aef43803dd/volumes" Feb 20 06:56:53 crc kubenswrapper[4492]: I0220 06:56:53.662093 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-65f5db898d-bxntp"] Feb 20 06:56:53 crc kubenswrapper[4492]: I0220 06:56:53.683980 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"1968b24edc23a5de4ecabc4dcc18c03f061c8965af1f7e77792ae0d80771d5c6"} Feb 20 06:56:53 crc kubenswrapper[4492]: W0220 06:56:53.688881 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62194cf_bf99_4a79_9d4e_fa7a0f88b8a7.slice/crio-8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac WatchSource:0}: Error finding container 8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac: Status 404 returned error can't find the container with id 8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac Feb 20 06:56:53 crc kubenswrapper[4492]: E0220 06:56:53.714444 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:f0473f3e18dd17d7021c02e991298923\\\"\"" pod="openstack/cinder-db-sync-qqhvc" podUID="82ed151d-241b-465d-8b5d-d97fe51015df" Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.068021 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.210233 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-blbr6"] Feb 20 06:56:54 crc kubenswrapper[4492]: W0220 06:56:54.225866 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda83f66a1_3f7a_4ff0_bb95_3eff19f92019.slice/crio-af852c6dbfb0f5fbcd7406d0d78127cc305e24b19251e58ff62f4ae97f96de0b WatchSource:0}: Error finding container af852c6dbfb0f5fbcd7406d0d78127cc305e24b19251e58ff62f4ae97f96de0b: Status 404 returned error can't find the container with id af852c6dbfb0f5fbcd7406d0d78127cc305e24b19251e58ff62f4ae97f96de0b Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.225889 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75dfbfdb9-5flr6"] Feb 20 06:56:54 crc kubenswrapper[4492]: W0220 06:56:54.237690 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22416222_576a_4b0c_a659_145ec65738e3.slice/crio-dc40efc160010c20d2ca248fb41862005e002c64b43e3d5e98f0bb026c59e56c WatchSource:0}: Error finding container dc40efc160010c20d2ca248fb41862005e002c64b43e3d5e98f0bb026c59e56c: Status 404 returned error can't find the container with id dc40efc160010c20d2ca248fb41862005e002c64b43e3d5e98f0bb026c59e56c Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.245389 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.251669 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d96664ccd-bltcr"] Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.276439 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7pcm9"] Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.308785 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.382719 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cf7fbb848-5c8g6"] Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.488764 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66cc4998b5-ts9pl"] Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.739485 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gjtqr" event={"ID":"5b719757-21ff-4967-a478-287c671d2bc7","Type":"ContainerStarted","Data":"713a78e6ee4d4e6d901fc9c60d99ba58b4619f1476820ee39be63f437b86d1d1"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.742745 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf7fbb848-5c8g6" event={"ID":"3a568bef-11fc-479e-bf90-4c2030405335","Type":"ContainerStarted","Data":"33a80f7ce5836d764d9ad30d5388677e957e6d3468b59372fe4257a99f18aff1"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.746601 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" event={"ID":"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4","Type":"ContainerStarted","Data":"8b13377be4dd9894076f9c8340425075f2f433d38dcd926fcdbaa484279cf923"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.749093 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-blbr6" event={"ID":"a83f66a1-3f7a-4ff0-bb95-3eff19f92019","Type":"ContainerStarted","Data":"10a248028db76ad41a2932cdc1b43904df882eec0fcdf41f9a8c10691b38f7b5"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.749114 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-blbr6" event={"ID":"a83f66a1-3f7a-4ff0-bb95-3eff19f92019","Type":"ContainerStarted","Data":"af852c6dbfb0f5fbcd7406d0d78127cc305e24b19251e58ff62f4ae97f96de0b"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.751500 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c","Type":"ContainerStarted","Data":"2f55e468a9755e594a5bf58a1ba4df34b9254777a803d2c139cf1c4d00d28fcb"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.755877 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-s246x" event={"ID":"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc","Type":"ContainerStarted","Data":"266d6cf492d9d445c03d3eeb2f23bcd26684b3384081232720c9f7aa0fd70dc6"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.770195 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65f5db898d-bxntp" event={"ID":"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7","Type":"ContainerStarted","Data":"8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.770927 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gjtqr" podStartSLOduration=5.521705241 podStartE2EDuration="42.770912477s" podCreationTimestamp="2026-02-20 06:56:12 +0000 UTC" firstStartedPulling="2026-02-20 06:56:14.67112984 +0000 UTC m=+931.442418818" lastFinishedPulling="2026-02-20 06:56:51.920337076 +0000 UTC m=+968.691626054" observedRunningTime="2026-02-20 06:56:54.760539817 +0000 UTC m=+971.531828784" watchObservedRunningTime="2026-02-20 06:56:54.770912477 +0000 UTC m=+971.542201454" Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.772092 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4","Type":"ContainerStarted","Data":"5562b30dbda106563d7e147cb358b9aa362388fd25544e9d1aa2ea8c4ff6c318"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.777907 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d96664ccd-bltcr" event={"ID":"22416222-576a-4b0c-a659-145ec65738e3","Type":"ContainerStarted","Data":"dc40efc160010c20d2ca248fb41862005e002c64b43e3d5e98f0bb026c59e56c"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.779244 4492 generic.go:334] "Generic (PLEG): container finished" podID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerID="9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8" exitCode=0 Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.788273 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pcm9" event={"ID":"3ed937a1-e598-4418-81f6-cbbd240e4d4a","Type":"ContainerDied","Data":"9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.788315 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pcm9" event={"ID":"3ed937a1-e598-4418-81f6-cbbd240e4d4a","Type":"ContainerStarted","Data":"fb77c7cac457fee555a8f11ca3fe822d8842192ce7cf9c521bcfb14adee3afc1"} Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.810042 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-blbr6" podStartSLOduration=16.810016697000002 podStartE2EDuration="16.810016697s" podCreationTimestamp="2026-02-20 06:56:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:54.772312726 +0000 UTC m=+971.543601704" watchObservedRunningTime="2026-02-20 06:56:54.810016697 +0000 UTC m=+971.581305664" Feb 20 06:56:54 crc kubenswrapper[4492]: I0220 06:56:54.823336 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-s246x" podStartSLOduration=5.683833351 podStartE2EDuration="43.823314298s" podCreationTimestamp="2026-02-20 06:56:11 +0000 UTC" firstStartedPulling="2026-02-20 06:56:13.780978188 +0000 UTC m=+930.552267166" lastFinishedPulling="2026-02-20 06:56:51.920459136 +0000 UTC m=+968.691748113" observedRunningTime="2026-02-20 06:56:54.787086219 +0000 UTC m=+971.558375197" watchObservedRunningTime="2026-02-20 06:56:54.823314298 +0000 UTC m=+971.594603266" Feb 20 06:56:54 crc kubenswrapper[4492]: W0220 06:56:54.930944 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d6a439_bc05_40d2_92ed_2aae95cdd9fe.slice/crio-8fb5b1fdbacc35b693cd87fc7f7f6d71778a7c4107bcd739c6fed134d8ab90cd WatchSource:0}: Error finding container 8fb5b1fdbacc35b693cd87fc7f7f6d71778a7c4107bcd739c6fed134d8ab90cd: Status 404 returned error can't find the container with id 8fb5b1fdbacc35b693cd87fc7f7f6d71778a7c4107bcd739c6fed134d8ab90cd Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.811385 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65f5db898d-bxntp" event={"ID":"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7","Type":"ContainerStarted","Data":"953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.820212 4492 generic.go:334] "Generic (PLEG): container finished" podID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerID="137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983" exitCode=0 Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.820355 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" event={"ID":"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4","Type":"ContainerDied","Data":"137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.843293 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d96664ccd-bltcr" event={"ID":"22416222-576a-4b0c-a659-145ec65738e3","Type":"ContainerStarted","Data":"aab8e586565f997c04b1e050051a9e1e90ff3e757c69053cec772aac6f0d3e49"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.863357 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cc4998b5-ts9pl" event={"ID":"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe","Type":"ContainerStarted","Data":"adc5d853b4c31761759209632306c75f5851927f5ec19ce42ec17cddccfaff94"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.863404 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cc4998b5-ts9pl" event={"ID":"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe","Type":"ContainerStarted","Data":"8fb5b1fdbacc35b693cd87fc7f7f6d71778a7c4107bcd739c6fed134d8ab90cd"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.865039 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c","Type":"ContainerStarted","Data":"791d37251c916b5ad0955620326747ec7ba4e7f3bee0bfa94d5f44b2e9d79f7b"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.874079 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerStarted","Data":"28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.887943 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf7fbb848-5c8g6" event={"ID":"3a568bef-11fc-479e-bf90-4c2030405335","Type":"ContainerStarted","Data":"f8b0f7955b6291ba776eb73ad0cea1f5f6d1f8b4ce4d244cf6a302da792fbc0f"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.887971 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf7fbb848-5c8g6" event={"ID":"3a568bef-11fc-479e-bf90-4c2030405335","Type":"ContainerStarted","Data":"b351db4079da572e45b8ead2fa564d38656849c6e0ce66ea7c1fbcbdb0671376"} Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.888994 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:56:55 crc kubenswrapper[4492]: I0220 06:56:55.923554 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cf7fbb848-5c8g6" podStartSLOduration=13.923539688 podStartE2EDuration="13.923539688s" podCreationTimestamp="2026-02-20 06:56:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:55.913851859 +0000 UTC m=+972.685140826" watchObservedRunningTime="2026-02-20 06:56:55.923539688 +0000 UTC m=+972.694828686" Feb 20 06:56:56 crc kubenswrapper[4492]: I0220 06:56:56.902677 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65f5db898d-bxntp" event={"ID":"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7","Type":"ContainerStarted","Data":"dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec"} Feb 20 06:56:56 crc kubenswrapper[4492]: I0220 06:56:56.904856 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4","Type":"ContainerStarted","Data":"3531f06088ea1446421176cebc399f5f56668f4f5bcf32653d49a885b6e93120"} Feb 20 06:56:56 crc kubenswrapper[4492]: I0220 06:56:56.910367 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pcm9" event={"ID":"3ed937a1-e598-4418-81f6-cbbd240e4d4a","Type":"ContainerStarted","Data":"8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d"} Feb 20 06:56:56 crc kubenswrapper[4492]: I0220 06:56:56.918550 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c","Type":"ContainerStarted","Data":"d823a7bd255c0ae804fce0a370bccf1720339bbda06fb122791a412db0e2cdf8"} Feb 20 06:56:56 crc kubenswrapper[4492]: I0220 06:56:56.935102 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-log" containerID="cri-o://791d37251c916b5ad0955620326747ec7ba4e7f3bee0bfa94d5f44b2e9d79f7b" gracePeriod=30 Feb 20 06:56:56 crc kubenswrapper[4492]: I0220 06:56:56.935289 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-httpd" containerID="cri-o://d823a7bd255c0ae804fce0a370bccf1720339bbda06fb122791a412db0e2cdf8" gracePeriod=30 Feb 20 06:56:56 crc kubenswrapper[4492]: I0220 06:56:56.971071 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-65f5db898d-bxntp" podStartSLOduration=35.552536085 podStartE2EDuration="36.971042209s" podCreationTimestamp="2026-02-20 06:56:20 +0000 UTC" firstStartedPulling="2026-02-20 06:56:53.701372921 +0000 UTC m=+970.472661899" lastFinishedPulling="2026-02-20 06:56:55.119879045 +0000 UTC m=+971.891168023" observedRunningTime="2026-02-20 06:56:56.93512725 +0000 UTC m=+973.706416228" watchObservedRunningTime="2026-02-20 06:56:56.971042209 +0000 UTC m=+973.742331187" Feb 20 06:56:57 crc kubenswrapper[4492]: I0220 06:56:57.067685 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=37.067661869 podStartE2EDuration="37.067661869s" podCreationTimestamp="2026-02-20 06:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:57.062750502 +0000 UTC m=+973.834039490" watchObservedRunningTime="2026-02-20 06:56:57.067661869 +0000 UTC m=+973.838950847" Feb 20 06:56:57 crc kubenswrapper[4492]: I0220 06:56:57.940803 4492 generic.go:334] "Generic (PLEG): container finished" podID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerID="791d37251c916b5ad0955620326747ec7ba4e7f3bee0bfa94d5f44b2e9d79f7b" exitCode=143 Feb 20 06:56:57 crc kubenswrapper[4492]: I0220 06:56:57.941008 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c","Type":"ContainerDied","Data":"791d37251c916b5ad0955620326747ec7ba4e7f3bee0bfa94d5f44b2e9d79f7b"} Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.961675 4492 generic.go:334] "Generic (PLEG): container finished" podID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerID="d823a7bd255c0ae804fce0a370bccf1720339bbda06fb122791a412db0e2cdf8" exitCode=0 Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.961955 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c","Type":"ContainerDied","Data":"d823a7bd255c0ae804fce0a370bccf1720339bbda06fb122791a412db0e2cdf8"} Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.966289 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" event={"ID":"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4","Type":"ContainerStarted","Data":"eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac"} Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.967358 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.977715 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4","Type":"ContainerStarted","Data":"6c7dc3ec381f94105dfea12471a974944ff297f7673a3edc346945980974bcac"} Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.978008 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-log" containerID="cri-o://3531f06088ea1446421176cebc399f5f56668f4f5bcf32653d49a885b6e93120" gracePeriod=30 Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.978605 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-httpd" containerID="cri-o://6c7dc3ec381f94105dfea12471a974944ff297f7673a3edc346945980974bcac" gracePeriod=30 Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.981621 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d96664ccd-bltcr" event={"ID":"22416222-576a-4b0c-a659-145ec65738e3","Type":"ContainerStarted","Data":"f09ba4baf1f6d76269afa712bb0c34d0c82b23215a47f548f6cf2c6479690712"} Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.989438 4492 generic.go:334] "Generic (PLEG): container finished" podID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerID="8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d" exitCode=0 Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.989562 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pcm9" event={"ID":"3ed937a1-e598-4418-81f6-cbbd240e4d4a","Type":"ContainerDied","Data":"8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d"} Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.991036 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" podStartSLOduration=16.991022775 podStartE2EDuration="16.991022775s" podCreationTimestamp="2026-02-20 06:56:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:58.98282279 +0000 UTC m=+975.754111768" watchObservedRunningTime="2026-02-20 06:56:58.991022775 +0000 UTC m=+975.762311752" Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.995218 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cc4998b5-ts9pl" event={"ID":"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe","Type":"ContainerStarted","Data":"1c818445e17f3731ac053e7ae72dbebcde628e1d2f0c41055bd747bf8a71e1fa"} Feb 20 06:56:58 crc kubenswrapper[4492]: I0220 06:56:58.995423 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:56:59 crc kubenswrapper[4492]: I0220 06:56:59.007415 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=39.007403928 podStartE2EDuration="39.007403928s" podCreationTimestamp="2026-02-20 06:56:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:59.000589114 +0000 UTC m=+975.771878102" watchObservedRunningTime="2026-02-20 06:56:59.007403928 +0000 UTC m=+975.778692906" Feb 20 06:56:59 crc kubenswrapper[4492]: I0220 06:56:59.033565 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d96664ccd-bltcr" podStartSLOduration=37.153563022 podStartE2EDuration="38.033548974s" podCreationTimestamp="2026-02-20 06:56:21 +0000 UTC" firstStartedPulling="2026-02-20 06:56:54.256089938 +0000 UTC m=+971.027378917" lastFinishedPulling="2026-02-20 06:56:55.136075891 +0000 UTC m=+971.907364869" observedRunningTime="2026-02-20 06:56:59.025809157 +0000 UTC m=+975.797098135" watchObservedRunningTime="2026-02-20 06:56:59.033548974 +0000 UTC m=+975.804837952" Feb 20 06:56:59 crc kubenswrapper[4492]: I0220 06:56:59.045528 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66cc4998b5-ts9pl" podStartSLOduration=15.045508193 podStartE2EDuration="15.045508193s" podCreationTimestamp="2026-02-20 06:56:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:56:59.044278026 +0000 UTC m=+975.815567003" watchObservedRunningTime="2026-02-20 06:56:59.045508193 +0000 UTC m=+975.816797172" Feb 20 06:57:00 crc kubenswrapper[4492]: I0220 06:57:00.030722 4492 generic.go:334] "Generic (PLEG): container finished" podID="5b719757-21ff-4967-a478-287c671d2bc7" containerID="713a78e6ee4d4e6d901fc9c60d99ba58b4619f1476820ee39be63f437b86d1d1" exitCode=0 Feb 20 06:57:00 crc kubenswrapper[4492]: I0220 06:57:00.030888 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gjtqr" event={"ID":"5b719757-21ff-4967-a478-287c671d2bc7","Type":"ContainerDied","Data":"713a78e6ee4d4e6d901fc9c60d99ba58b4619f1476820ee39be63f437b86d1d1"} Feb 20 06:57:00 crc kubenswrapper[4492]: I0220 06:57:00.038849 4492 generic.go:334] "Generic (PLEG): container finished" podID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerID="6c7dc3ec381f94105dfea12471a974944ff297f7673a3edc346945980974bcac" exitCode=143 Feb 20 06:57:00 crc kubenswrapper[4492]: I0220 06:57:00.038901 4492 generic.go:334] "Generic (PLEG): container finished" podID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerID="3531f06088ea1446421176cebc399f5f56668f4f5bcf32653d49a885b6e93120" exitCode=143 Feb 20 06:57:00 crc kubenswrapper[4492]: I0220 06:57:00.039076 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4","Type":"ContainerDied","Data":"6c7dc3ec381f94105dfea12471a974944ff297f7673a3edc346945980974bcac"} Feb 20 06:57:00 crc kubenswrapper[4492]: I0220 06:57:00.039185 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4","Type":"ContainerDied","Data":"3531f06088ea1446421176cebc399f5f56668f4f5bcf32653d49a885b6e93120"} Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.074808 4492 generic.go:334] "Generic (PLEG): container finished" podID="7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" containerID="266d6cf492d9d445c03d3eeb2f23bcd26684b3384081232720c9f7aa0fd70dc6" exitCode=0 Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.075132 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-s246x" event={"ID":"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc","Type":"ContainerDied","Data":"266d6cf492d9d445c03d3eeb2f23bcd26684b3384081232720c9f7aa0fd70dc6"} Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.084370 4492 generic.go:334] "Generic (PLEG): container finished" podID="a83f66a1-3f7a-4ff0-bb95-3eff19f92019" containerID="10a248028db76ad41a2932cdc1b43904df882eec0fcdf41f9a8c10691b38f7b5" exitCode=0 Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.084410 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-blbr6" event={"ID":"a83f66a1-3f7a-4ff0-bb95-3eff19f92019","Type":"ContainerDied","Data":"10a248028db76ad41a2932cdc1b43904df882eec0fcdf41f9a8c10691b38f7b5"} Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.268572 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.272601 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.418948 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:57:01 crc kubenswrapper[4492]: I0220 06:57:01.419820 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.010946 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.030677 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-scripts\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.030760 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.081852 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-scripts" (OuterVolumeSpecName: "scripts") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.083714 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.120082 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.120677 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4","Type":"ContainerDied","Data":"5562b30dbda106563d7e147cb358b9aa362388fd25544e9d1aa2ea8c4ff6c318"} Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.120735 4492 scope.go:117] "RemoveContainer" containerID="6c7dc3ec381f94105dfea12471a974944ff297f7673a3edc346945980974bcac" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.133919 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-public-tls-certs\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.133964 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-config-data\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.134035 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-httpd-run\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.134138 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-logs\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.134246 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ppk9\" (UniqueName: \"kubernetes.io/projected/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-kube-api-access-6ppk9\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.134788 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-combined-ca-bundle\") pod \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\" (UID: \"a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4\") " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.135552 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.135593 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.136952 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-logs" (OuterVolumeSpecName: "logs") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.137110 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.160462 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-kube-api-access-6ppk9" (OuterVolumeSpecName: "kube-api-access-6ppk9") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "kube-api-access-6ppk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.178994 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.192498 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.204622 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.233718 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-config-data" (OuterVolumeSpecName: "config-data") pod "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" (UID: "a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.238453 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.238488 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.238499 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.238509 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.238518 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.238527 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ppk9\" (UniqueName: \"kubernetes.io/projected/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-kube-api-access-6ppk9\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.238537 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.478754 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.504518 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.512080 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:57:02 crc kubenswrapper[4492]: E0220 06:57:02.512543 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-log" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.512563 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-log" Feb 20 06:57:02 crc kubenswrapper[4492]: E0220 06:57:02.512600 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-httpd" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.512608 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-httpd" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.512826 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-log" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.512853 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" containerName="glance-httpd" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.514319 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.520708 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.520887 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.532940 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.661090 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-logs\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.661193 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.661229 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.661271 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.661313 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb9nt\" (UniqueName: \"kubernetes.io/projected/6b76c1c3-dda9-4648-a623-83d73d2168dc-kube-api-access-qb9nt\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.661447 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.661488 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.662082 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765418 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765535 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765587 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb9nt\" (UniqueName: \"kubernetes.io/projected/6b76c1c3-dda9-4648-a623-83d73d2168dc-kube-api-access-qb9nt\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765708 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765740 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765785 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765869 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-logs\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.765930 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.766312 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.766844 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.767462 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-logs\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.773206 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.777286 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.779613 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-config-data\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.780139 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-scripts\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.785035 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb9nt\" (UniqueName: \"kubernetes.io/projected/6b76c1c3-dda9-4648-a623-83d73d2168dc-kube-api-access-qb9nt\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.799006 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " pod="openstack/glance-default-external-api-0" Feb 20 06:57:02 crc kubenswrapper[4492]: I0220 06:57:02.855027 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:57:03 crc kubenswrapper[4492]: I0220 06:57:03.063658 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:57:03 crc kubenswrapper[4492]: I0220 06:57:03.116541 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-596d946f85-qdrjh"] Feb 20 06:57:03 crc kubenswrapper[4492]: I0220 06:57:03.117584 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerName="dnsmasq-dns" containerID="cri-o://35dbf0337debb0422acca4c4dbbd7060833037e4bbe0444696d53a3569b9eb2d" gracePeriod=10 Feb 20 06:57:03 crc kubenswrapper[4492]: I0220 06:57:03.343000 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Feb 20 06:57:03 crc kubenswrapper[4492]: I0220 06:57:03.565467 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4" path="/var/lib/kubelet/pods/a6b172cd-42ea-4e97-b4b0-e053dcbe3ff4/volumes" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.154167 4492 generic.go:334] "Generic (PLEG): container finished" podID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerID="35dbf0337debb0422acca4c4dbbd7060833037e4bbe0444696d53a3569b9eb2d" exitCode=0 Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.154380 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" event={"ID":"9a1f3e39-8b98-4bce-ae51-fa786e589c02","Type":"ContainerDied","Data":"35dbf0337debb0422acca4c4dbbd7060833037e4bbe0444696d53a3569b9eb2d"} Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.751292 4492 scope.go:117] "RemoveContainer" containerID="3531f06088ea1446421176cebc399f5f56668f4f5bcf32653d49a885b6e93120" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.805813 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.806404 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gjtqr" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.807981 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.809642 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-s246x" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908041 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-combined-ca-bundle\") pod \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908351 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-credential-keys\") pod \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908431 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4b2qh\" (UniqueName: \"kubernetes.io/projected/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-kube-api-access-4b2qh\") pod \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908540 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-httpd-run\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908619 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-config-data\") pod \"5b719757-21ff-4967-a478-287c671d2bc7\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908708 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-config-data\") pod \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908789 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-combined-ca-bundle\") pod \"5b719757-21ff-4967-a478-287c671d2bc7\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.908878 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-internal-tls-certs\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909026 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2v56\" (UniqueName: \"kubernetes.io/projected/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-kube-api-access-p2v56\") pod \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\" (UID: \"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909095 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fllcl\" (UniqueName: \"kubernetes.io/projected/5b719757-21ff-4967-a478-287c671d2bc7-kube-api-access-fllcl\") pod \"5b719757-21ff-4967-a478-287c671d2bc7\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909157 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-logs\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909262 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-scripts\") pod \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909335 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909397 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-config-data\") pod \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909485 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-combined-ca-bundle\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909583 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czf7l\" (UniqueName: \"kubernetes.io/projected/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-kube-api-access-czf7l\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909665 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-config-data\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909747 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-scripts\") pod \"5b719757-21ff-4967-a478-287c671d2bc7\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909802 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-combined-ca-bundle\") pod \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909861 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b719757-21ff-4967-a478-287c671d2bc7-logs\") pod \"5b719757-21ff-4967-a478-287c671d2bc7\" (UID: \"5b719757-21ff-4967-a478-287c671d2bc7\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.909934 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-fernet-keys\") pod \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\" (UID: \"a83f66a1-3f7a-4ff0-bb95-3eff19f92019\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.912626 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-scripts\") pod \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\" (UID: \"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c\") " Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.918727 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-scripts" (OuterVolumeSpecName: "scripts") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.926840 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a83f66a1-3f7a-4ff0-bb95-3eff19f92019" (UID: "a83f66a1-3f7a-4ff0-bb95-3eff19f92019"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.933668 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-scripts" (OuterVolumeSpecName: "scripts") pod "a83f66a1-3f7a-4ff0-bb95-3eff19f92019" (UID: "a83f66a1-3f7a-4ff0-bb95-3eff19f92019"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.945666 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-logs" (OuterVolumeSpecName: "logs") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.946277 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-kube-api-access-4b2qh" (OuterVolumeSpecName: "kube-api-access-4b2qh") pod "a83f66a1-3f7a-4ff0-bb95-3eff19f92019" (UID: "a83f66a1-3f7a-4ff0-bb95-3eff19f92019"). InnerVolumeSpecName "kube-api-access-4b2qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.946395 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.954312 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.955706 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-kube-api-access-p2v56" (OuterVolumeSpecName: "kube-api-access-p2v56") pod "7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" (UID: "7ee3d5af-cee2-4398-b8c8-969d7aa69fdc"). InnerVolumeSpecName "kube-api-access-p2v56". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.956100 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b719757-21ff-4967-a478-287c671d2bc7-logs" (OuterVolumeSpecName: "logs") pod "5b719757-21ff-4967-a478-287c671d2bc7" (UID: "5b719757-21ff-4967-a478-287c671d2bc7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.959705 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-scripts" (OuterVolumeSpecName: "scripts") pod "5b719757-21ff-4967-a478-287c671d2bc7" (UID: "5b719757-21ff-4967-a478-287c671d2bc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.970590 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b719757-21ff-4967-a478-287c671d2bc7-kube-api-access-fllcl" (OuterVolumeSpecName: "kube-api-access-fllcl") pod "5b719757-21ff-4967-a478-287c671d2bc7" (UID: "5b719757-21ff-4967-a478-287c671d2bc7"). InnerVolumeSpecName "kube-api-access-fllcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:04 crc kubenswrapper[4492]: I0220 06:57:04.976579 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-kube-api-access-czf7l" (OuterVolumeSpecName: "kube-api-access-czf7l") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "kube-api-access-czf7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.006234 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a83f66a1-3f7a-4ff0-bb95-3eff19f92019" (UID: "a83f66a1-3f7a-4ff0-bb95-3eff19f92019"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015046 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2v56\" (UniqueName: \"kubernetes.io/projected/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-kube-api-access-p2v56\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015084 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fllcl\" (UniqueName: \"kubernetes.io/projected/5b719757-21ff-4967-a478-287c671d2bc7-kube-api-access-fllcl\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015094 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015104 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015134 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015144 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czf7l\" (UniqueName: \"kubernetes.io/projected/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-kube-api-access-czf7l\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015155 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015165 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b719757-21ff-4967-a478-287c671d2bc7-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015173 4492 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015180 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015189 4492 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015199 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4b2qh\" (UniqueName: \"kubernetes.io/projected/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-kube-api-access-4b2qh\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.015211 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.044618 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a83f66a1-3f7a-4ff0-bb95-3eff19f92019" (UID: "a83f66a1-3f7a-4ff0-bb95-3eff19f92019"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.048316 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" (UID: "7ee3d5af-cee2-4398-b8c8-969d7aa69fdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.064072 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-config-data" (OuterVolumeSpecName: "config-data") pod "5b719757-21ff-4967-a478-287c671d2bc7" (UID: "5b719757-21ff-4967-a478-287c671d2bc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.093761 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-config-data" (OuterVolumeSpecName: "config-data") pod "a83f66a1-3f7a-4ff0-bb95-3eff19f92019" (UID: "a83f66a1-3f7a-4ff0-bb95-3eff19f92019"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.201077 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.201112 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83f66a1-3f7a-4ff0-bb95-3eff19f92019-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.201125 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.201141 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.311998 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-blbr6" event={"ID":"a83f66a1-3f7a-4ff0-bb95-3eff19f92019","Type":"ContainerDied","Data":"af852c6dbfb0f5fbcd7406d0d78127cc305e24b19251e58ff62f4ae97f96de0b"} Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.312045 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af852c6dbfb0f5fbcd7406d0d78127cc305e24b19251e58ff62f4ae97f96de0b" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.312115 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-blbr6" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.313563 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b719757-21ff-4967-a478-287c671d2bc7" (UID: "5b719757-21ff-4967-a478-287c671d2bc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.324562 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.324565 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c","Type":"ContainerDied","Data":"2f55e468a9755e594a5bf58a1ba4df34b9254777a803d2c139cf1c4d00d28fcb"} Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.324668 4492 scope.go:117] "RemoveContainer" containerID="d823a7bd255c0ae804fce0a370bccf1720339bbda06fb122791a412db0e2cdf8" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.324869 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.328336 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.339078 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-s246x" event={"ID":"7ee3d5af-cee2-4398-b8c8-969d7aa69fdc","Type":"ContainerDied","Data":"6121ddd77210f62d523ad6eb8cfb464a2641e64d6a21479ef7e06c553c32a9ca"} Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.339119 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6121ddd77210f62d523ad6eb8cfb464a2641e64d6a21479ef7e06c553c32a9ca" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.339199 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-s246x" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.342745 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gjtqr" event={"ID":"5b719757-21ff-4967-a478-287c671d2bc7","Type":"ContainerDied","Data":"8ef0971b3fe24ea6d3b15ef40c8a9a0fd722bdd0b512fb31fe9f7352aedb7252"} Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.342786 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ef0971b3fe24ea6d3b15ef40c8a9a0fd722bdd0b512fb31fe9f7352aedb7252" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.342858 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gjtqr" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.374050 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.381082 4492 scope.go:117] "RemoveContainer" containerID="791d37251c916b5ad0955620326747ec7ba4e7f3bee0bfa94d5f44b2e9d79f7b" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.414843 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-nb\") pod \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.432156 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-sb\") pod \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.432349 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-config\") pod \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.432409 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmslp\" (UniqueName: \"kubernetes.io/projected/9a1f3e39-8b98-4bce-ae51-fa786e589c02-kube-api-access-vmslp\") pod \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.430266 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.433051 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-swift-storage-0\") pod \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.433132 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-svc\") pod \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\" (UID: \"9a1f3e39-8b98-4bce-ae51-fa786e589c02\") " Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.434206 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b719757-21ff-4967-a478-287c671d2bc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.434221 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.434230 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.434240 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.453028 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a1f3e39-8b98-4bce-ae51-fa786e589c02-kube-api-access-vmslp" (OuterVolumeSpecName: "kube-api-access-vmslp") pod "9a1f3e39-8b98-4bce-ae51-fa786e589c02" (UID: "9a1f3e39-8b98-4bce-ae51-fa786e589c02"). InnerVolumeSpecName "kube-api-access-vmslp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.527107 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-config-data" (OuterVolumeSpecName: "config-data") pod "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" (UID: "87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.536863 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmslp\" (UniqueName: \"kubernetes.io/projected/9a1f3e39-8b98-4bce-ae51-fa786e589c02-kube-api-access-vmslp\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.536889 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.538200 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9a1f3e39-8b98-4bce-ae51-fa786e589c02" (UID: "9a1f3e39-8b98-4bce-ae51-fa786e589c02"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.545118 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a1f3e39-8b98-4bce-ae51-fa786e589c02" (UID: "9a1f3e39-8b98-4bce-ae51-fa786e589c02"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.573353 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a1f3e39-8b98-4bce-ae51-fa786e589c02" (UID: "9a1f3e39-8b98-4bce-ae51-fa786e589c02"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.613437 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a1f3e39-8b98-4bce-ae51-fa786e589c02" (UID: "9a1f3e39-8b98-4bce-ae51-fa786e589c02"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.613745 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-config-data" (OuterVolumeSpecName: "config-data") pod "7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" (UID: "7ee3d5af-cee2-4398-b8c8-969d7aa69fdc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.629933 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-config" (OuterVolumeSpecName: "config") pod "9a1f3e39-8b98-4bce-ae51-fa786e589c02" (UID: "9a1f3e39-8b98-4bce-ae51-fa786e589c02"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.631327 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:57:05 crc kubenswrapper[4492]: W0220 06:57:05.636695 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b76c1c3_dda9_4648_a623_83d73d2168dc.slice/crio-606bd7f86e5b76be302aa4240e78cf39a3c2e3423318d2daf2a24c733560ac0d WatchSource:0}: Error finding container 606bd7f86e5b76be302aa4240e78cf39a3c2e3423318d2daf2a24c733560ac0d: Status 404 returned error can't find the container with id 606bd7f86e5b76be302aa4240e78cf39a3c2e3423318d2daf2a24c733560ac0d Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.638120 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.638144 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.638153 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.638161 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.638171 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.638178 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a1f3e39-8b98-4bce-ae51-fa786e589c02-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.666586 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.689578 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698246 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:57:05 crc kubenswrapper[4492]: E0220 06:57:05.698721 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" containerName="heat-db-sync" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698750 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" containerName="heat-db-sync" Feb 20 06:57:05 crc kubenswrapper[4492]: E0220 06:57:05.698776 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b719757-21ff-4967-a478-287c671d2bc7" containerName="placement-db-sync" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698783 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b719757-21ff-4967-a478-287c671d2bc7" containerName="placement-db-sync" Feb 20 06:57:05 crc kubenswrapper[4492]: E0220 06:57:05.698799 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerName="dnsmasq-dns" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698807 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerName="dnsmasq-dns" Feb 20 06:57:05 crc kubenswrapper[4492]: E0220 06:57:05.698815 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83f66a1-3f7a-4ff0-bb95-3eff19f92019" containerName="keystone-bootstrap" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698821 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83f66a1-3f7a-4ff0-bb95-3eff19f92019" containerName="keystone-bootstrap" Feb 20 06:57:05 crc kubenswrapper[4492]: E0220 06:57:05.698829 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerName="init" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698834 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerName="init" Feb 20 06:57:05 crc kubenswrapper[4492]: E0220 06:57:05.698852 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-log" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698858 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-log" Feb 20 06:57:05 crc kubenswrapper[4492]: E0220 06:57:05.698864 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-httpd" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.698870 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-httpd" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.699077 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83f66a1-3f7a-4ff0-bb95-3eff19f92019" containerName="keystone-bootstrap" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.699114 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" containerName="dnsmasq-dns" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.699127 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-log" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.699142 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" containerName="glance-httpd" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.699162 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" containerName="heat-db-sync" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.699177 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b719757-21ff-4967-a478-287c671d2bc7" containerName="placement-db-sync" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.700180 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.708372 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.709073 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.724976 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843272 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843676 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843728 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-logs\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843769 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843839 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5np9\" (UniqueName: \"kubernetes.io/projected/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-kube-api-access-v5np9\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843879 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843907 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.843942 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950076 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5np9\" (UniqueName: \"kubernetes.io/projected/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-kube-api-access-v5np9\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950186 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950228 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950293 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950345 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950376 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950451 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-logs\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950541 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.950988 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.963339 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.964131 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-logs\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.971033 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.971306 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.974352 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.987383 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:05 crc kubenswrapper[4492]: I0220 06:57:05.992358 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.001188 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5np9\" (UniqueName: \"kubernetes.io/projected/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-kube-api-access-v5np9\") pod \"glance-default-internal-api-0\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.026595 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-565b895f9d-w6nzp"] Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.039091 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.044209 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.044406 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rr7tt" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.044561 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.045901 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.046026 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.047332 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-565b895f9d-w6nzp"] Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.098395 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.121356 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7fcbbc74fb-jdvms"] Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.124219 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.137180 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.137339 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4v2r8" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.137426 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.137663 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.137713 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.137670 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.162929 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcg6p\" (UniqueName: \"kubernetes.io/projected/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-kube-api-access-dcg6p\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.163032 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-combined-ca-bundle\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.163103 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-public-tls-certs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.163129 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-public-tls-certs\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.163177 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-scripts\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.163260 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-credential-keys\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.163442 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-fernet-keys\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.163465 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-scripts\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.189989 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-config-data\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.190040 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfzbr\" (UniqueName: \"kubernetes.io/projected/87b134e6-2d14-4fc6-bfd9-6055300dab55-kube-api-access-gfzbr\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.190067 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-config-data\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.190118 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-internal-tls-certs\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.190136 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-combined-ca-bundle\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.190155 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b134e6-2d14-4fc6-bfd9-6055300dab55-logs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.190181 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-internal-tls-certs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.190261 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fcbbc74fb-jdvms"] Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312232 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-fernet-keys\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312612 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-scripts\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312675 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-config-data\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312707 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfzbr\" (UniqueName: \"kubernetes.io/projected/87b134e6-2d14-4fc6-bfd9-6055300dab55-kube-api-access-gfzbr\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312740 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-config-data\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312780 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-internal-tls-certs\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312793 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-combined-ca-bundle\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312813 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b134e6-2d14-4fc6-bfd9-6055300dab55-logs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312833 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-internal-tls-certs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312863 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcg6p\" (UniqueName: \"kubernetes.io/projected/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-kube-api-access-dcg6p\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312910 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-combined-ca-bundle\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312951 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-public-tls-certs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312966 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-public-tls-certs\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.312999 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-scripts\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.313046 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-credential-keys\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.332260 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-credential-keys\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.332303 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-fernet-keys\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.332514 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b134e6-2d14-4fc6-bfd9-6055300dab55-logs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.345222 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-combined-ca-bundle\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.350683 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-public-tls-certs\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.351377 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-internal-tls-certs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.358780 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-internal-tls-certs\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.359234 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-config-data\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.372773 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfzbr\" (UniqueName: \"kubernetes.io/projected/87b134e6-2d14-4fc6-bfd9-6055300dab55-kube-api-access-gfzbr\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.398327 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-scripts\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.406730 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcg6p\" (UniqueName: \"kubernetes.io/projected/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-kube-api-access-dcg6p\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.427349 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d79gm" event={"ID":"05565eb0-90db-4ec7-b646-db528fca40ad","Type":"ContainerStarted","Data":"5243939417ea46c026a621e02fc03fc327f97535c2337c2c6f66159db20dd056"} Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.449312 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-public-tls-certs\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.451426 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c87d9c3-ee9a-430b-a943-e92ea78e65a5-scripts\") pod \"keystone-7fcbbc74fb-jdvms\" (UID: \"4c87d9c3-ee9a-430b-a943-e92ea78e65a5\") " pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.455901 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-combined-ca-bundle\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.457867 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-config-data\") pod \"placement-565b895f9d-w6nzp\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.472080 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.474023 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerStarted","Data":"516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032"} Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.487640 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-d79gm" podStartSLOduration=3.800348792 podStartE2EDuration="54.487618226s" podCreationTimestamp="2026-02-20 06:56:12 +0000 UTC" firstStartedPulling="2026-02-20 06:56:14.236399082 +0000 UTC m=+931.007688059" lastFinishedPulling="2026-02-20 06:57:04.923668525 +0000 UTC m=+981.694957493" observedRunningTime="2026-02-20 06:57:06.470187596 +0000 UTC m=+983.241476574" watchObservedRunningTime="2026-02-20 06:57:06.487618226 +0000 UTC m=+983.258907205" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.520772 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" event={"ID":"9a1f3e39-8b98-4bce-ae51-fa786e589c02","Type":"ContainerDied","Data":"9675dc1eaf331252194c53f96bc0186656187594fd0fe60bafab9449d98f3c0a"} Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.525577 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pcm9" event={"ID":"3ed937a1-e598-4418-81f6-cbbd240e4d4a","Type":"ContainerStarted","Data":"50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8"} Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.521700 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-596d946f85-qdrjh" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.528595 4492 scope.go:117] "RemoveContainer" containerID="35dbf0337debb0422acca4c4dbbd7060833037e4bbe0444696d53a3569b9eb2d" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.545702 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b76c1c3-dda9-4648-a623-83d73d2168dc","Type":"ContainerStarted","Data":"606bd7f86e5b76be302aa4240e78cf39a3c2e3423318d2daf2a24c733560ac0d"} Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.575097 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7pcm9" podStartSLOduration=19.615818969 podStartE2EDuration="29.575074135s" podCreationTimestamp="2026-02-20 06:56:37 +0000 UTC" firstStartedPulling="2026-02-20 06:56:54.920830398 +0000 UTC m=+971.692119376" lastFinishedPulling="2026-02-20 06:57:04.880085564 +0000 UTC m=+981.651374542" observedRunningTime="2026-02-20 06:57:06.567845451 +0000 UTC m=+983.339134430" watchObservedRunningTime="2026-02-20 06:57:06.575074135 +0000 UTC m=+983.346363114" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.609754 4492 scope.go:117] "RemoveContainer" containerID="e60b90c3bc6755170de6a3c42f38dfe317bf1783d765457ec558f41b2a5ddf34" Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.641603 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-596d946f85-qdrjh"] Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.701522 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-596d946f85-qdrjh"] Feb 20 06:57:06 crc kubenswrapper[4492]: I0220 06:57:06.704196 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.015035 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.171536 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fcbbc74fb-jdvms"] Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.401925 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-565b895f9d-w6nzp"] Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.458962 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.459941 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:57:07 crc kubenswrapper[4492]: W0220 06:57:07.485276 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87b134e6_2d14_4fc6_bfd9_6055300dab55.slice/crio-1c7bd0523be39fc059a7d34142485b485e8b0a35ff121c7f30ff0143702c239b WatchSource:0}: Error finding container 1c7bd0523be39fc059a7d34142485b485e8b0a35ff121c7f30ff0143702c239b: Status 404 returned error can't find the container with id 1c7bd0523be39fc059a7d34142485b485e8b0a35ff121c7f30ff0143702c239b Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.613892 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c" path="/var/lib/kubelet/pods/87fbd5c6-b5dd-4747-bcf0-f9bd33701d8c/volumes" Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.614890 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a1f3e39-8b98-4bce-ae51-fa786e589c02" path="/var/lib/kubelet/pods/9a1f3e39-8b98-4bce-ae51-fa786e589c02/volumes" Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.657409 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fcbbc74fb-jdvms" event={"ID":"4c87d9c3-ee9a-430b-a943-e92ea78e65a5","Type":"ContainerStarted","Data":"a3edba63af866746d997820f976e1c0a38b72ce2634aee1733a133bbee6078bf"} Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.713904 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b76c1c3-dda9-4648-a623-83d73d2168dc","Type":"ContainerStarted","Data":"d383ad6877a1cecb15de4177e5a8f4e75cae890d3e67c1c0fecf31b099218869"} Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.728088 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a","Type":"ContainerStarted","Data":"e7a13de06402e79c32a7a10172ee37d601bb02e0106ac7f8801ede6e80e1a8c4"} Feb 20 06:57:07 crc kubenswrapper[4492]: I0220 06:57:07.730110 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565b895f9d-w6nzp" event={"ID":"87b134e6-2d14-4fc6-bfd9-6055300dab55","Type":"ContainerStarted","Data":"1c7bd0523be39fc059a7d34142485b485e8b0a35ff121c7f30ff0143702c239b"} Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.292206 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-65bf8dfc4b-nwxp9"] Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.293843 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.321631 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65bf8dfc4b-nwxp9"] Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.427453 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-internal-tls-certs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.427793 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-scripts\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.427816 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-config-data\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.427851 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b944r\" (UniqueName: \"kubernetes.io/projected/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-kube-api-access-b944r\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.427917 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-public-tls-certs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.427954 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-logs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.427994 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-combined-ca-bundle\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.530105 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-public-tls-certs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.530194 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-logs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.530261 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-combined-ca-bundle\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.530284 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-internal-tls-certs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.530337 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-scripts\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.530359 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-config-data\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.530383 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b944r\" (UniqueName: \"kubernetes.io/projected/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-kube-api-access-b944r\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.533758 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-logs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.537318 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-scripts\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.537757 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-combined-ca-bundle\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.539300 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-config-data\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.540863 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-internal-tls-certs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.545227 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b944r\" (UniqueName: \"kubernetes.io/projected/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-kube-api-access-b944r\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.551864 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eacd0a76-6ee4-42ea-b2b8-7faef8c27f68-public-tls-certs\") pod \"placement-65bf8dfc4b-nwxp9\" (UID: \"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68\") " pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.628578 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7pcm9" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="registry-server" probeResult="failure" output=< Feb 20 06:57:08 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 06:57:08 crc kubenswrapper[4492]: > Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.679173 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.810349 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fcbbc74fb-jdvms" event={"ID":"4c87d9c3-ee9a-430b-a943-e92ea78e65a5","Type":"ContainerStarted","Data":"3909e989b72f8b534f13ea743619d4fa69653dac4c8c6d0aa11ea4592588bb2b"} Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.810977 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.839775 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b76c1c3-dda9-4648-a623-83d73d2168dc","Type":"ContainerStarted","Data":"fbe8afa1ed5c0e0186d700ca9b8ac3b53341dc72899bf4679af6c90d025a3f82"} Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.896547 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a","Type":"ContainerStarted","Data":"273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747"} Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.904674 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7fcbbc74fb-jdvms" podStartSLOduration=2.90463576 podStartE2EDuration="2.90463576s" podCreationTimestamp="2026-02-20 06:57:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:08.840144994 +0000 UTC m=+985.611433972" watchObservedRunningTime="2026-02-20 06:57:08.90463576 +0000 UTC m=+985.675924738" Feb 20 06:57:08 crc kubenswrapper[4492]: I0220 06:57:08.920743 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565b895f9d-w6nzp" event={"ID":"87b134e6-2d14-4fc6-bfd9-6055300dab55","Type":"ContainerStarted","Data":"78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2"} Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.198687 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.198668985 podStartE2EDuration="7.198668985s" podCreationTimestamp="2026-02-20 06:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:08.911744617 +0000 UTC m=+985.683033596" watchObservedRunningTime="2026-02-20 06:57:09.198668985 +0000 UTC m=+985.969957963" Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.207548 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65bf8dfc4b-nwxp9"] Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.935884 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a","Type":"ContainerStarted","Data":"6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08"} Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.957798 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565b895f9d-w6nzp" event={"ID":"87b134e6-2d14-4fc6-bfd9-6055300dab55","Type":"ContainerStarted","Data":"7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199"} Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.958591 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.959720 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.975985 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.975971861 podStartE2EDuration="4.975971861s" podCreationTimestamp="2026-02-20 06:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:09.971852588 +0000 UTC m=+986.743141565" watchObservedRunningTime="2026-02-20 06:57:09.975971861 +0000 UTC m=+986.747260839" Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.978362 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65bf8dfc4b-nwxp9" event={"ID":"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68","Type":"ContainerStarted","Data":"3b6fa9e6a372081fd2660b2db98a98c464c02c5d45d8c61349aeaeb76f7d4b14"} Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.978388 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.978398 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65bf8dfc4b-nwxp9" event={"ID":"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68","Type":"ContainerStarted","Data":"5b4495220202316ff704fb0c9264358f17727682e1f2b39ed63a21dcd85da27e"} Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.978752 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:09 crc kubenswrapper[4492]: I0220 06:57:09.996289 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-565b895f9d-w6nzp" podStartSLOduration=4.996281469 podStartE2EDuration="4.996281469s" podCreationTimestamp="2026-02-20 06:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:09.990198444 +0000 UTC m=+986.761487422" watchObservedRunningTime="2026-02-20 06:57:09.996281469 +0000 UTC m=+986.767570446" Feb 20 06:57:10 crc kubenswrapper[4492]: I0220 06:57:10.019870 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-65bf8dfc4b-nwxp9" podStartSLOduration=2.019846612 podStartE2EDuration="2.019846612s" podCreationTimestamp="2026-02-20 06:57:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:10.010601968 +0000 UTC m=+986.781890947" watchObservedRunningTime="2026-02-20 06:57:10.019846612 +0000 UTC m=+986.791135590" Feb 20 06:57:10 crc kubenswrapper[4492]: I0220 06:57:10.989776 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65bf8dfc4b-nwxp9" event={"ID":"eacd0a76-6ee4-42ea-b2b8-7faef8c27f68","Type":"ContainerStarted","Data":"9b515b6e02987642ab9874119182a99241ade3afdd5893808451fe377b2eafe6"} Feb 20 06:57:11 crc kubenswrapper[4492]: I0220 06:57:10.996676 4492 generic.go:334] "Generic (PLEG): container finished" podID="05565eb0-90db-4ec7-b646-db528fca40ad" containerID="5243939417ea46c026a621e02fc03fc327f97535c2337c2c6f66159db20dd056" exitCode=0 Feb 20 06:57:11 crc kubenswrapper[4492]: I0220 06:57:11.000706 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d79gm" event={"ID":"05565eb0-90db-4ec7-b646-db528fca40ad","Type":"ContainerDied","Data":"5243939417ea46c026a621e02fc03fc327f97535c2337c2c6f66159db20dd056"} Feb 20 06:57:11 crc kubenswrapper[4492]: I0220 06:57:11.010562 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qqhvc" event={"ID":"82ed151d-241b-465d-8b5d-d97fe51015df","Type":"ContainerStarted","Data":"9dff1702bfc3db58505276fe1faa25079b995045b18caee73490500f3e3d21b5"} Feb 20 06:57:11 crc kubenswrapper[4492]: I0220 06:57:11.046085 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-qqhvc" podStartSLOduration=5.363739048 podStartE2EDuration="1m0.046070149s" podCreationTimestamp="2026-02-20 06:56:11 +0000 UTC" firstStartedPulling="2026-02-20 06:56:13.761695697 +0000 UTC m=+930.532984675" lastFinishedPulling="2026-02-20 06:57:08.444026798 +0000 UTC m=+985.215315776" observedRunningTime="2026-02-20 06:57:11.03430218 +0000 UTC m=+987.805591158" watchObservedRunningTime="2026-02-20 06:57:11.046070149 +0000 UTC m=+987.817359117" Feb 20 06:57:11 crc kubenswrapper[4492]: I0220 06:57:11.268508 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65f5db898d-bxntp" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 20 06:57:11 crc kubenswrapper[4492]: I0220 06:57:11.421664 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d96664ccd-bltcr" podUID="22416222-576a-4b0c-a659-145ec65738e3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 20 06:57:12 crc kubenswrapper[4492]: I0220 06:57:12.856800 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 06:57:12 crc kubenswrapper[4492]: I0220 06:57:12.857201 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 06:57:12 crc kubenswrapper[4492]: I0220 06:57:12.905255 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 06:57:12 crc kubenswrapper[4492]: I0220 06:57:12.914303 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.051140 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.051189 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.328736 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.577311 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66cc4998b5-ts9pl"] Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.577530 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66cc4998b5-ts9pl" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-api" containerID="cri-o://adc5d853b4c31761759209632306c75f5851927f5ec19ce42ec17cddccfaff94" gracePeriod=30 Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.577678 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66cc4998b5-ts9pl" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-httpd" containerID="cri-o://1c818445e17f3731ac053e7ae72dbebcde628e1d2f0c41055bd747bf8a71e1fa" gracePeriod=30 Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.606464 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-65749686ff-sv9sq"] Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.611573 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.616286 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-66cc4998b5-ts9pl" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": EOF" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.632575 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65749686ff-sv9sq"] Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.764931 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-ovndb-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.765017 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-internal-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.765039 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-config\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.765057 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-httpd-config\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.765098 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99zcp\" (UniqueName: \"kubernetes.io/projected/495833af-a575-4097-9033-e968a74f3a18-kube-api-access-99zcp\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.765130 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-public-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.765145 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-combined-ca-bundle\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.867041 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-httpd-config\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.867122 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99zcp\" (UniqueName: \"kubernetes.io/projected/495833af-a575-4097-9033-e968a74f3a18-kube-api-access-99zcp\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.867162 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-public-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.867181 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-combined-ca-bundle\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.867265 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-ovndb-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.867339 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-internal-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.867363 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-config\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.876305 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-internal-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.877398 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-ovndb-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.877669 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-httpd-config\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.877804 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-combined-ca-bundle\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.884883 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-public-tls-certs\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.894187 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-config\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.897744 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99zcp\" (UniqueName: \"kubernetes.io/projected/495833af-a575-4097-9033-e968a74f3a18-kube-api-access-99zcp\") pod \"neutron-65749686ff-sv9sq\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:13 crc kubenswrapper[4492]: I0220 06:57:13.928042 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:14 crc kubenswrapper[4492]: I0220 06:57:14.086040 4492 generic.go:334] "Generic (PLEG): container finished" podID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerID="1c818445e17f3731ac053e7ae72dbebcde628e1d2f0c41055bd747bf8a71e1fa" exitCode=0 Feb 20 06:57:14 crc kubenswrapper[4492]: I0220 06:57:14.086625 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cc4998b5-ts9pl" event={"ID":"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe","Type":"ContainerDied","Data":"1c818445e17f3731ac053e7ae72dbebcde628e1d2f0c41055bd747bf8a71e1fa"} Feb 20 06:57:15 crc kubenswrapper[4492]: I0220 06:57:15.094727 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-66cc4998b5-ts9pl" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": dial tcp 10.217.0.157:9696: connect: connection refused" Feb 20 06:57:15 crc kubenswrapper[4492]: I0220 06:57:15.103253 4492 generic.go:334] "Generic (PLEG): container finished" podID="82ed151d-241b-465d-8b5d-d97fe51015df" containerID="9dff1702bfc3db58505276fe1faa25079b995045b18caee73490500f3e3d21b5" exitCode=0 Feb 20 06:57:15 crc kubenswrapper[4492]: I0220 06:57:15.103311 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qqhvc" event={"ID":"82ed151d-241b-465d-8b5d-d97fe51015df","Type":"ContainerDied","Data":"9dff1702bfc3db58505276fe1faa25079b995045b18caee73490500f3e3d21b5"} Feb 20 06:57:15 crc kubenswrapper[4492]: I0220 06:57:15.996096 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 06:57:15 crc kubenswrapper[4492]: I0220 06:57:15.996232 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:57:15 crc kubenswrapper[4492]: I0220 06:57:15.997198 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 06:57:16 crc kubenswrapper[4492]: I0220 06:57:16.106677 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:16 crc kubenswrapper[4492]: I0220 06:57:16.106971 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:16 crc kubenswrapper[4492]: I0220 06:57:16.151605 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:16 crc kubenswrapper[4492]: I0220 06:57:16.167064 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.019259 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d79gm" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.142182 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-d79gm" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.142232 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-d79gm" event={"ID":"05565eb0-90db-4ec7-b646-db528fca40ad","Type":"ContainerDied","Data":"ec2315be0308f36c13788159dace69c10acd2d1b387f2ee8678d5ba7269fe415"} Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.142264 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec2315be0308f36c13788159dace69c10acd2d1b387f2ee8678d5ba7269fe415" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.144296 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.144325 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.150140 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-db-sync-config-data\") pod \"05565eb0-90db-4ec7-b646-db528fca40ad\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.150253 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z57n\" (UniqueName: \"kubernetes.io/projected/05565eb0-90db-4ec7-b646-db528fca40ad-kube-api-access-5z57n\") pod \"05565eb0-90db-4ec7-b646-db528fca40ad\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.150325 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-combined-ca-bundle\") pod \"05565eb0-90db-4ec7-b646-db528fca40ad\" (UID: \"05565eb0-90db-4ec7-b646-db528fca40ad\") " Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.157921 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "05565eb0-90db-4ec7-b646-db528fca40ad" (UID: "05565eb0-90db-4ec7-b646-db528fca40ad"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.202914 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05565eb0-90db-4ec7-b646-db528fca40ad" (UID: "05565eb0-90db-4ec7-b646-db528fca40ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.217729 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05565eb0-90db-4ec7-b646-db528fca40ad-kube-api-access-5z57n" (OuterVolumeSpecName: "kube-api-access-5z57n") pod "05565eb0-90db-4ec7-b646-db528fca40ad" (UID: "05565eb0-90db-4ec7-b646-db528fca40ad"). InnerVolumeSpecName "kube-api-access-5z57n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.258013 4492 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.258614 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z57n\" (UniqueName: \"kubernetes.io/projected/05565eb0-90db-4ec7-b646-db528fca40ad-kube-api-access-5z57n\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:17 crc kubenswrapper[4492]: I0220 06:57:17.258663 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05565eb0-90db-4ec7-b646-db528fca40ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.315292 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-f5cc678f7-42jkm"] Feb 20 06:57:18 crc kubenswrapper[4492]: E0220 06:57:18.315967 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05565eb0-90db-4ec7-b646-db528fca40ad" containerName="barbican-db-sync" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.315982 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="05565eb0-90db-4ec7-b646-db528fca40ad" containerName="barbican-db-sync" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.316181 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="05565eb0-90db-4ec7-b646-db528fca40ad" containerName="barbican-db-sync" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.317169 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.332726 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.332971 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fslbj" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.333191 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.359761 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f5cc678f7-42jkm"] Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.376817 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.389920 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-config-data-custom\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.390024 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3f17e6-f16a-46db-a73c-331307b540cd-logs\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.390074 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-combined-ca-bundle\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.390170 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbxt6\" (UniqueName: \"kubernetes.io/projected/2d3f17e6-f16a-46db-a73c-331307b540cd-kube-api-access-tbxt6\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.390339 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-config-data\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.419877 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7b75bfd866-zx65r"] Feb 20 06:57:18 crc kubenswrapper[4492]: E0220 06:57:18.420431 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ed151d-241b-465d-8b5d-d97fe51015df" containerName="cinder-db-sync" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.420454 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ed151d-241b-465d-8b5d-d97fe51015df" containerName="cinder-db-sync" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.420674 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ed151d-241b-465d-8b5d-d97fe51015df" containerName="cinder-db-sync" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.421773 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.428891 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.481862 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7987756ddf-ppjxp"] Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.486815 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.492240 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-config-data\") pod \"82ed151d-241b-465d-8b5d-d97fe51015df\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.496030 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-db-sync-config-data\") pod \"82ed151d-241b-465d-8b5d-d97fe51015df\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.496214 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ed151d-241b-465d-8b5d-d97fe51015df-etc-machine-id\") pod \"82ed151d-241b-465d-8b5d-d97fe51015df\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.496599 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxg4m\" (UniqueName: \"kubernetes.io/projected/82ed151d-241b-465d-8b5d-d97fe51015df-kube-api-access-vxg4m\") pod \"82ed151d-241b-465d-8b5d-d97fe51015df\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.496646 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-scripts\") pod \"82ed151d-241b-465d-8b5d-d97fe51015df\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.496697 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-combined-ca-bundle\") pod \"82ed151d-241b-465d-8b5d-d97fe51015df\" (UID: \"82ed151d-241b-465d-8b5d-d97fe51015df\") " Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497244 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-config-data\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497322 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-config-data\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497367 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-config-data-custom\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497489 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-config-data-custom\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497515 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-combined-ca-bundle\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497598 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjc7n\" (UniqueName: \"kubernetes.io/projected/fa37a18e-7d2d-45db-8988-0ec443201155-kube-api-access-wjc7n\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497620 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3f17e6-f16a-46db-a73c-331307b540cd-logs\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.497889 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-combined-ca-bundle\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.498066 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa37a18e-7d2d-45db-8988-0ec443201155-logs\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.498093 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbxt6\" (UniqueName: \"kubernetes.io/projected/2d3f17e6-f16a-46db-a73c-331307b540cd-kube-api-access-tbxt6\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.499359 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d3f17e6-f16a-46db-a73c-331307b540cd-logs\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.499543 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82ed151d-241b-465d-8b5d-d97fe51015df-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "82ed151d-241b-465d-8b5d-d97fe51015df" (UID: "82ed151d-241b-465d-8b5d-d97fe51015df"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.523186 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7987756ddf-ppjxp"] Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.534258 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-scripts" (OuterVolumeSpecName: "scripts") pod "82ed151d-241b-465d-8b5d-d97fe51015df" (UID: "82ed151d-241b-465d-8b5d-d97fe51015df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.547901 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b75bfd866-zx65r"] Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.548340 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ed151d-241b-465d-8b5d-d97fe51015df-kube-api-access-vxg4m" (OuterVolumeSpecName: "kube-api-access-vxg4m") pod "82ed151d-241b-465d-8b5d-d97fe51015df" (UID: "82ed151d-241b-465d-8b5d-d97fe51015df"). InnerVolumeSpecName "kube-api-access-vxg4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.548489 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7pcm9" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="registry-server" probeResult="failure" output=< Feb 20 06:57:18 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 06:57:18 crc kubenswrapper[4492]: > Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.552822 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-config-data-custom\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.554434 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-combined-ca-bundle\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.555957 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbxt6\" (UniqueName: \"kubernetes.io/projected/2d3f17e6-f16a-46db-a73c-331307b540cd-kube-api-access-tbxt6\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.559879 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d3f17e6-f16a-46db-a73c-331307b540cd-config-data\") pod \"barbican-worker-f5cc678f7-42jkm\" (UID: \"2d3f17e6-f16a-46db-a73c-331307b540cd\") " pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.561617 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "82ed151d-241b-465d-8b5d-d97fe51015df" (UID: "82ed151d-241b-465d-8b5d-d97fe51015df"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.599822 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-swift-storage-0\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.599884 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-config-data\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.599912 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-config-data-custom\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.599955 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-svc\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.599989 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-nb\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600016 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-combined-ca-bundle\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600064 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjc7n\" (UniqueName: \"kubernetes.io/projected/fa37a18e-7d2d-45db-8988-0ec443201155-kube-api-access-wjc7n\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600159 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-sb\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600181 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdl7v\" (UniqueName: \"kubernetes.io/projected/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-kube-api-access-rdl7v\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600210 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa37a18e-7d2d-45db-8988-0ec443201155-logs\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600254 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-config\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600334 4492 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/82ed151d-241b-465d-8b5d-d97fe51015df-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600352 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxg4m\" (UniqueName: \"kubernetes.io/projected/82ed151d-241b-465d-8b5d-d97fe51015df-kube-api-access-vxg4m\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600362 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.600371 4492 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.601710 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa37a18e-7d2d-45db-8988-0ec443201155-logs\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.613539 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-config-data-custom\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.614600 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-config-data\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.615620 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82ed151d-241b-465d-8b5d-d97fe51015df" (UID: "82ed151d-241b-465d-8b5d-d97fe51015df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.616024 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa37a18e-7d2d-45db-8988-0ec443201155-combined-ca-bundle\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.624763 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7547fb57c8-ndpch"] Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.626435 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.628370 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7547fb57c8-ndpch"] Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.630301 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.639579 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-config-data" (OuterVolumeSpecName: "config-data") pod "82ed151d-241b-465d-8b5d-d97fe51015df" (UID: "82ed151d-241b-465d-8b5d-d97fe51015df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.646534 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjc7n\" (UniqueName: \"kubernetes.io/projected/fa37a18e-7d2d-45db-8988-0ec443201155-kube-api-access-wjc7n\") pod \"barbican-keystone-listener-7b75bfd866-zx65r\" (UID: \"fa37a18e-7d2d-45db-8988-0ec443201155\") " pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.673713 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f5cc678f7-42jkm" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707715 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data-custom\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707766 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-config\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707819 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5de1b303-1a15-44d9-b3fd-dfc3381dc491-logs\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707853 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-swift-storage-0\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707872 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-combined-ca-bundle\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707890 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707927 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-svc\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.707954 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-nb\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.708008 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnz45\" (UniqueName: \"kubernetes.io/projected/5de1b303-1a15-44d9-b3fd-dfc3381dc491-kube-api-access-gnz45\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.708046 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-sb\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.708066 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdl7v\" (UniqueName: \"kubernetes.io/projected/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-kube-api-access-rdl7v\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.708120 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.708132 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ed151d-241b-465d-8b5d-d97fe51015df-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.709186 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-config\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.709711 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-svc\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.713599 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-swift-storage-0\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.715111 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-nb\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.718706 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-sb\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.725818 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdl7v\" (UniqueName: \"kubernetes.io/projected/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-kube-api-access-rdl7v\") pod \"dnsmasq-dns-7987756ddf-ppjxp\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.758561 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.811811 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data-custom\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.811883 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5de1b303-1a15-44d9-b3fd-dfc3381dc491-logs\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.812225 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-combined-ca-bundle\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.812255 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.812323 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnz45\" (UniqueName: \"kubernetes.io/projected/5de1b303-1a15-44d9-b3fd-dfc3381dc491-kube-api-access-gnz45\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.812390 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5de1b303-1a15-44d9-b3fd-dfc3381dc491-logs\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.816182 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data-custom\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.818665 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.818907 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-combined-ca-bundle\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.834344 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnz45\" (UniqueName: \"kubernetes.io/projected/5de1b303-1a15-44d9-b3fd-dfc3381dc491-kube-api-access-gnz45\") pod \"barbican-api-7547fb57c8-ndpch\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.958407 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:18 crc kubenswrapper[4492]: I0220 06:57:18.970501 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.185843 4492 generic.go:334] "Generic (PLEG): container finished" podID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerID="adc5d853b4c31761759209632306c75f5851927f5ec19ce42ec17cddccfaff94" exitCode=0 Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.185967 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cc4998b5-ts9pl" event={"ID":"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe","Type":"ContainerDied","Data":"adc5d853b4c31761759209632306c75f5851927f5ec19ce42ec17cddccfaff94"} Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.191614 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-qqhvc" event={"ID":"82ed151d-241b-465d-8b5d-d97fe51015df","Type":"ContainerDied","Data":"0ab7743ef909a89d2ed691be403a464f86185976b1dc84851ce3eddd12cfa308"} Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.191683 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ab7743ef909a89d2ed691be403a464f86185976b1dc84851ce3eddd12cfa308" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.191813 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-qqhvc" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.688237 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.690444 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.695215 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.695315 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.695670 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-bj5qz" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.695826 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.712770 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.752952 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-scripts\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.753072 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjfkt\" (UniqueName: \"kubernetes.io/projected/bd72cda4-24aa-4f03-ba43-767587bb7925-kube-api-access-rjfkt\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.753251 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.753314 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd72cda4-24aa-4f03-ba43-767587bb7925-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.753368 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.753440 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.778056 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987756ddf-ppjxp"] Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.812245 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d4fbd665-cbckt"] Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.813718 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.848393 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d4fbd665-cbckt"] Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.864791 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-sb\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865173 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865274 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd72cda4-24aa-4f03-ba43-767587bb7925-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865347 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-svc\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865374 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865401 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-config\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865721 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865777 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-scripts\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865802 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-nb\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.865895 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd72cda4-24aa-4f03-ba43-767587bb7925-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.880307 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.880536 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjfkt\" (UniqueName: \"kubernetes.io/projected/bd72cda4-24aa-4f03-ba43-767587bb7925-kube-api-access-rjfkt\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.880592 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-swift-storage-0\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.880709 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tcth\" (UniqueName: \"kubernetes.io/projected/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-kube-api-access-5tcth\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.887821 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.892868 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-scripts\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.916121 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjfkt\" (UniqueName: \"kubernetes.io/projected/bd72cda4-24aa-4f03-ba43-767587bb7925-kube-api-access-rjfkt\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.921220 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.982697 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tcth\" (UniqueName: \"kubernetes.io/projected/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-kube-api-access-5tcth\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.982792 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-sb\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.982919 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-svc\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.982955 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-config\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.983026 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-nb\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.983082 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-swift-storage-0\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.984051 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-swift-storage-0\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.984879 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-sb\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.991525 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-config\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:19 crc kubenswrapper[4492]: I0220 06:57:19.992138 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-nb\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.003685 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-svc\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.024026 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tcth\" (UniqueName: \"kubernetes.io/projected/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-kube-api-access-5tcth\") pod \"dnsmasq-dns-59d4fbd665-cbckt\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.024076 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.026050 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.034088 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.052879 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.080135 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.085046 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.085207 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-logs\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.085302 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.085398 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mgt2\" (UniqueName: \"kubernetes.io/projected/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-kube-api-access-5mgt2\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.085423 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.085455 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-scripts\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.085522 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.149634 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.187172 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mgt2\" (UniqueName: \"kubernetes.io/projected/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-kube-api-access-5mgt2\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.187230 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.187266 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-scripts\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.187314 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.187340 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.187457 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-logs\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.187561 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.194630 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-scripts\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.196675 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-logs\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.196911 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.201100 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.202413 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.209002 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data-custom\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.229035 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mgt2\" (UniqueName: \"kubernetes.io/projected/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-kube-api-access-5mgt2\") pod \"cinder-api-0\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.369306 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.443809 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f5cc678f7-42jkm"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.477691 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.477821 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:57:20 crc kubenswrapper[4492]: E0220 06:57:20.482561 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.514533 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.609370 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-config\") pod \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.610768 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-ovndb-tls-certs\") pod \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.610883 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-internal-tls-certs\") pod \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.612466 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-httpd-config\") pod \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.619225 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-public-tls-certs\") pod \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.619528 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-combined-ca-bundle\") pod \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.619625 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnl9q\" (UniqueName: \"kubernetes.io/projected/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-kube-api-access-fnl9q\") pod \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\" (UID: \"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe\") " Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.647717 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" (UID: "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.651307 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987756ddf-ppjxp"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.662763 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-kube-api-access-fnl9q" (OuterVolumeSpecName: "kube-api-access-fnl9q") pod "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" (UID: "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe"). InnerVolumeSpecName "kube-api-access-fnl9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.665082 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7547fb57c8-ndpch"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.672427 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7b75bfd866-zx65r"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.723610 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.723643 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnl9q\" (UniqueName: \"kubernetes.io/projected/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-kube-api-access-fnl9q\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.751177 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65749686ff-sv9sq"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.763036 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-config" (OuterVolumeSpecName: "config") pod "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" (UID: "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.826896 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.856607 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" (UID: "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.859490 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" (UID: "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.893838 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.908389 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.935631 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" (UID: "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.937176 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.937200 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.937209 4492 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.963859 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d4fbd665-cbckt"] Feb 20 06:57:20 crc kubenswrapper[4492]: I0220 06:57:20.981308 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" (UID: "b6d6a439-bc05-40d2-92ed-2aae95cdd9fe"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.042900 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.087494 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:21 crc kubenswrapper[4492]: W0220 06:57:21.119500 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod504bb2ee_6c5c_4858_ba94_88ccc1267e5d.slice/crio-de4315f67f7a989ea29398a40cb76f302696a4d57bcf8ffdb1aae92a6c7f57c1 WatchSource:0}: Error finding container de4315f67f7a989ea29398a40cb76f302696a4d57bcf8ffdb1aae92a6c7f57c1: Status 404 returned error can't find the container with id de4315f67f7a989ea29398a40cb76f302696a4d57bcf8ffdb1aae92a6c7f57c1 Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.235394 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7547fb57c8-ndpch" event={"ID":"5de1b303-1a15-44d9-b3fd-dfc3381dc491","Type":"ContainerStarted","Data":"f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.236229 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7547fb57c8-ndpch" event={"ID":"5de1b303-1a15-44d9-b3fd-dfc3381dc491","Type":"ContainerStarted","Data":"439bbc88010ae80f02e83a26cda2fa2dbcd75f9755d99945cc10255b89df4020"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.238307 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" event={"ID":"fa37a18e-7d2d-45db-8988-0ec443201155","Type":"ContainerStarted","Data":"4247774434c967153443670581a00fdb4ad2397eae2971a91d4335938efc279f"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.239596 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" event={"ID":"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b","Type":"ContainerStarted","Data":"04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.239638 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" event={"ID":"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b","Type":"ContainerStarted","Data":"a12fd2303bc2aa68838d3f9d66caf57e3afdf5a0744f54f32562520f9f564a1f"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.243724 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f5cc678f7-42jkm" event={"ID":"2d3f17e6-f16a-46db-a73c-331307b540cd","Type":"ContainerStarted","Data":"85ef29c8e0099762c55d35f2e9bcf7d18c62b5a31f0f9d99b613495fd1d2ff3e"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.250386 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerStarted","Data":"64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.250706 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="ceilometer-notification-agent" containerID="cri-o://28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f" gracePeriod=30 Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.250986 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.255623 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="sg-core" containerID="cri-o://516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032" gracePeriod=30 Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.255736 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="proxy-httpd" containerID="cri-o://64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb" gracePeriod=30 Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.261446 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" event={"ID":"8bdbc0d0-444f-495f-bd72-57c41c6bcd71","Type":"ContainerStarted","Data":"6592d07ab3200040ce1380a5185adb60d9d213deaf37653b2f5645919867d4a9"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.267518 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-65f5db898d-bxntp" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.317154 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65749686ff-sv9sq" event={"ID":"495833af-a575-4097-9033-e968a74f3a18","Type":"ContainerStarted","Data":"022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.317552 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65749686ff-sv9sq" event={"ID":"495833af-a575-4097-9033-e968a74f3a18","Type":"ContainerStarted","Data":"f8851eb0435e46350e5644f98d9320468afe894129862ee6e917b2c442bb4669"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.360600 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cc4998b5-ts9pl" event={"ID":"b6d6a439-bc05-40d2-92ed-2aae95cdd9fe","Type":"ContainerDied","Data":"8fb5b1fdbacc35b693cd87fc7f7f6d71778a7c4107bcd739c6fed134d8ab90cd"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.360719 4492 scope.go:117] "RemoveContainer" containerID="1c818445e17f3731ac053e7ae72dbebcde628e1d2f0c41055bd747bf8a71e1fa" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.360987 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cc4998b5-ts9pl" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.385673 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd72cda4-24aa-4f03-ba43-767587bb7925","Type":"ContainerStarted","Data":"aa427d55678ecbce2f55d192d58b3291f9fea1a3ad7600f566944170a9c2bc68"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.395937 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"504bb2ee-6c5c-4858-ba94-88ccc1267e5d","Type":"ContainerStarted","Data":"de4315f67f7a989ea29398a40cb76f302696a4d57bcf8ffdb1aae92a6c7f57c1"} Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.419247 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d96664ccd-bltcr" podUID="22416222-576a-4b0c-a659-145ec65738e3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.535426 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66cc4998b5-ts9pl"] Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.539145 4492 scope.go:117] "RemoveContainer" containerID="adc5d853b4c31761759209632306c75f5851927f5ec19ce42ec17cddccfaff94" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.545879 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66cc4998b5-ts9pl"] Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.625638 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" path="/var/lib/kubelet/pods/b6d6a439-bc05-40d2-92ed-2aae95cdd9fe/volumes" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.754025 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.881162 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-sb\") pod \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.881509 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdl7v\" (UniqueName: \"kubernetes.io/projected/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-kube-api-access-rdl7v\") pod \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.881628 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-config\") pod \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.881690 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-nb\") pod \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.881747 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-swift-storage-0\") pod \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.881844 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-svc\") pod \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\" (UID: \"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b\") " Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.931345 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" (UID: "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.966783 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-kube-api-access-rdl7v" (OuterVolumeSpecName: "kube-api-access-rdl7v") pod "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" (UID: "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b"). InnerVolumeSpecName "kube-api-access-rdl7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.972731 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" (UID: "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.981770 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-config" (OuterVolumeSpecName: "config") pod "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" (UID: "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.984925 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.984950 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdl7v\" (UniqueName: \"kubernetes.io/projected/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-kube-api-access-rdl7v\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.984962 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:21 crc kubenswrapper[4492]: I0220 06:57:21.984972 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.016103 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" (UID: "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.016686 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" (UID: "fb8bf48a-3c9d-40c7-ba6b-abfbea18730b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.086451 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.086498 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.429142 4492 generic.go:334] "Generic (PLEG): container finished" podID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerID="06beb6d9fd8852edc6ee30c9dde953a2b48c8c56c5915464977e08077d7ef9d1" exitCode=0 Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.429207 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" event={"ID":"8bdbc0d0-444f-495f-bd72-57c41c6bcd71","Type":"ContainerDied","Data":"06beb6d9fd8852edc6ee30c9dde953a2b48c8c56c5915464977e08077d7ef9d1"} Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.441843 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65749686ff-sv9sq" event={"ID":"495833af-a575-4097-9033-e968a74f3a18","Type":"ContainerStarted","Data":"b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b"} Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.442890 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.465739 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.491159 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7547fb57c8-ndpch" event={"ID":"5de1b303-1a15-44d9-b3fd-dfc3381dc491","Type":"ContainerStarted","Data":"b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c"} Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.492749 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.492781 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.511843 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-65749686ff-sv9sq" podStartSLOduration=9.511826444 podStartE2EDuration="9.511826444s" podCreationTimestamp="2026-02-20 06:57:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:22.497755485 +0000 UTC m=+999.269044453" watchObservedRunningTime="2026-02-20 06:57:22.511826444 +0000 UTC m=+999.283115422" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.529065 4492 generic.go:334] "Generic (PLEG): container finished" podID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerID="64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb" exitCode=0 Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.529102 4492 generic.go:334] "Generic (PLEG): container finished" podID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerID="516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032" exitCode=2 Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.529196 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerDied","Data":"64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb"} Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.529244 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerDied","Data":"516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032"} Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.529901 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7547fb57c8-ndpch" podStartSLOduration=4.529877626 podStartE2EDuration="4.529877626s" podCreationTimestamp="2026-02-20 06:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:22.525302393 +0000 UTC m=+999.296591371" watchObservedRunningTime="2026-02-20 06:57:22.529877626 +0000 UTC m=+999.301166604" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.531274 4492 generic.go:334] "Generic (PLEG): container finished" podID="fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" containerID="04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1" exitCode=0 Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.531316 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" event={"ID":"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b","Type":"ContainerDied","Data":"04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1"} Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.531342 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" event={"ID":"fb8bf48a-3c9d-40c7-ba6b-abfbea18730b","Type":"ContainerDied","Data":"a12fd2303bc2aa68838d3f9d66caf57e3afdf5a0744f54f32562520f9f564a1f"} Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.531359 4492 scope.go:117] "RemoveContainer" containerID="04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.531455 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7987756ddf-ppjxp" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.592808 4492 scope.go:117] "RemoveContainer" containerID="04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1" Feb 20 06:57:22 crc kubenswrapper[4492]: E0220 06:57:22.593618 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1\": container with ID starting with 04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1 not found: ID does not exist" containerID="04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.593669 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1"} err="failed to get container status \"04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1\": rpc error: code = NotFound desc = could not find container \"04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1\": container with ID starting with 04231c8c4777dc12db0e4af92eaa1036b56d45dc372ea0ae2dcb092009b55fc1 not found: ID does not exist" Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.631124 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7987756ddf-ppjxp"] Feb 20 06:57:22 crc kubenswrapper[4492]: I0220 06:57:22.640671 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7987756ddf-ppjxp"] Feb 20 06:57:23 crc kubenswrapper[4492]: I0220 06:57:23.579794 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" path="/var/lib/kubelet/pods/fb8bf48a-3c9d-40c7-ba6b-abfbea18730b/volumes" Feb 20 06:57:23 crc kubenswrapper[4492]: I0220 06:57:23.586393 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" event={"ID":"8bdbc0d0-444f-495f-bd72-57c41c6bcd71","Type":"ContainerStarted","Data":"2ce5643fc51d0db7366057f9d56f5d38c433a025228ec00329e5e4047cc3d2f7"} Feb 20 06:57:23 crc kubenswrapper[4492]: I0220 06:57:23.586617 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd72cda4-24aa-4f03-ba43-767587bb7925","Type":"ContainerStarted","Data":"df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018"} Feb 20 06:57:23 crc kubenswrapper[4492]: I0220 06:57:23.586752 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"504bb2ee-6c5c-4858-ba94-88ccc1267e5d","Type":"ContainerStarted","Data":"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a"} Feb 20 06:57:23 crc kubenswrapper[4492]: I0220 06:57:23.692595 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" podStartSLOduration=4.69257731 podStartE2EDuration="4.69257731s" podCreationTimestamp="2026-02-20 06:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:23.683606544 +0000 UTC m=+1000.454895521" watchObservedRunningTime="2026-02-20 06:57:23.69257731 +0000 UTC m=+1000.463866288" Feb 20 06:57:24 crc kubenswrapper[4492]: I0220 06:57:24.664990 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f5cc678f7-42jkm" event={"ID":"2d3f17e6-f16a-46db-a73c-331307b540cd","Type":"ContainerStarted","Data":"2a1c50df23a351965249ac40a4b2bcbee55cc44e3c121242709b757c52dfb9a1"} Feb 20 06:57:24 crc kubenswrapper[4492]: I0220 06:57:24.670246 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"504bb2ee-6c5c-4858-ba94-88ccc1267e5d","Type":"ContainerStarted","Data":"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18"} Feb 20 06:57:24 crc kubenswrapper[4492]: I0220 06:57:24.670674 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api-log" containerID="cri-o://eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a" gracePeriod=30 Feb 20 06:57:24 crc kubenswrapper[4492]: I0220 06:57:24.670779 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 20 06:57:24 crc kubenswrapper[4492]: I0220 06:57:24.671239 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api" containerID="cri-o://a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18" gracePeriod=30 Feb 20 06:57:24 crc kubenswrapper[4492]: I0220 06:57:24.671539 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:24 crc kubenswrapper[4492]: I0220 06:57:24.701820 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.701802831 podStartE2EDuration="5.701802831s" podCreationTimestamp="2026-02-20 06:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:24.692562275 +0000 UTC m=+1001.463851263" watchObservedRunningTime="2026-02-20 06:57:24.701802831 +0000 UTC m=+1001.473091809" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.238438 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.299292 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-etc-machine-id\") pod \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.299338 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data-custom\") pod \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.299381 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mgt2\" (UniqueName: \"kubernetes.io/projected/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-kube-api-access-5mgt2\") pod \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.299405 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-combined-ca-bundle\") pod \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.299455 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-logs\") pod \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.299505 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-scripts\") pod \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.299544 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data\") pod \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\" (UID: \"504bb2ee-6c5c-4858-ba94-88ccc1267e5d\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.310452 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-kube-api-access-5mgt2" (OuterVolumeSpecName: "kube-api-access-5mgt2") pod "504bb2ee-6c5c-4858-ba94-88ccc1267e5d" (UID: "504bb2ee-6c5c-4858-ba94-88ccc1267e5d"). InnerVolumeSpecName "kube-api-access-5mgt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.310521 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "504bb2ee-6c5c-4858-ba94-88ccc1267e5d" (UID: "504bb2ee-6c5c-4858-ba94-88ccc1267e5d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.319298 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-logs" (OuterVolumeSpecName: "logs") pod "504bb2ee-6c5c-4858-ba94-88ccc1267e5d" (UID: "504bb2ee-6c5c-4858-ba94-88ccc1267e5d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.319906 4492 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.319926 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mgt2\" (UniqueName: \"kubernetes.io/projected/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-kube-api-access-5mgt2\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.319938 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.327524 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "504bb2ee-6c5c-4858-ba94-88ccc1267e5d" (UID: "504bb2ee-6c5c-4858-ba94-88ccc1267e5d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.335060 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-scripts" (OuterVolumeSpecName: "scripts") pod "504bb2ee-6c5c-4858-ba94-88ccc1267e5d" (UID: "504bb2ee-6c5c-4858-ba94-88ccc1267e5d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.376303 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data" (OuterVolumeSpecName: "config-data") pod "504bb2ee-6c5c-4858-ba94-88ccc1267e5d" (UID: "504bb2ee-6c5c-4858-ba94-88ccc1267e5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.377717 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "504bb2ee-6c5c-4858-ba94-88ccc1267e5d" (UID: "504bb2ee-6c5c-4858-ba94-88ccc1267e5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.422047 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.422079 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.422088 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.422099 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/504bb2ee-6c5c-4858-ba94-88ccc1267e5d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.429721 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9db58c4f4-5prd7"] Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.436988 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-httpd" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437012 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-httpd" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.437027 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437033 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.437050 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api-log" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437056 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api-log" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.437068 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" containerName="init" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437074 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" containerName="init" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.437087 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-api" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437092 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-api" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437246 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api-log" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437259 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-httpd" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437279 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d6a439-bc05-40d2-92ed-2aae95cdd9fe" containerName="neutron-api" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437286 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerName="cinder-api" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.437296 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb8bf48a-3c9d-40c7-ba6b-abfbea18730b" containerName="init" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.438267 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.448569 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.448797 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.454482 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9db58c4f4-5prd7"] Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.524032 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-config-data-custom\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.524095 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-combined-ca-bundle\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.524130 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwzkq\" (UniqueName: \"kubernetes.io/projected/4e3149f8-a4fe-484c-a0eb-9799994a144c-kube-api-access-wwzkq\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.524196 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-public-tls-certs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.524274 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-internal-tls-certs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.524363 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e3149f8-a4fe-484c-a0eb-9799994a144c-logs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.524451 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-config-data\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.626269 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-public-tls-certs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.626346 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-internal-tls-certs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.626383 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e3149f8-a4fe-484c-a0eb-9799994a144c-logs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.626421 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-config-data\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.626512 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-config-data-custom\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.626531 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-combined-ca-bundle\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.626547 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwzkq\" (UniqueName: \"kubernetes.io/projected/4e3149f8-a4fe-484c-a0eb-9799994a144c-kube-api-access-wwzkq\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.628838 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e3149f8-a4fe-484c-a0eb-9799994a144c-logs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.633384 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-internal-tls-certs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.634880 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-public-tls-certs\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.639128 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-combined-ca-bundle\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.639723 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-config-data\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.642278 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e3149f8-a4fe-484c-a0eb-9799994a144c-config-data-custom\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.645312 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwzkq\" (UniqueName: \"kubernetes.io/projected/4e3149f8-a4fe-484c-a0eb-9799994a144c-kube-api-access-wwzkq\") pod \"barbican-api-9db58c4f4-5prd7\" (UID: \"4e3149f8-a4fe-484c-a0eb-9799994a144c\") " pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.680582 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" event={"ID":"fa37a18e-7d2d-45db-8988-0ec443201155","Type":"ContainerStarted","Data":"dfe24558fa32f61960e434ad18b1d6de79c70582575f3f831026758497d8c97b"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.680637 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" event={"ID":"fa37a18e-7d2d-45db-8988-0ec443201155","Type":"ContainerStarted","Data":"2cb868d582283cdd0b91a7d0b5297bff0e6a017f27e1ba98b41b7ea4b2782a25"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.695779 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f5cc678f7-42jkm" event={"ID":"2d3f17e6-f16a-46db-a73c-331307b540cd","Type":"ContainerStarted","Data":"3370494dd050e2889c506773adad9a276c327741c925874bd23a3fbe418ed13d"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.705310 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7b75bfd866-zx65r" podStartSLOduration=4.086445809 podStartE2EDuration="7.705297823s" podCreationTimestamp="2026-02-20 06:57:18 +0000 UTC" firstStartedPulling="2026-02-20 06:57:20.701858471 +0000 UTC m=+997.473147449" lastFinishedPulling="2026-02-20 06:57:24.320710485 +0000 UTC m=+1001.091999463" observedRunningTime="2026-02-20 06:57:25.701757129 +0000 UTC m=+1002.473046108" watchObservedRunningTime="2026-02-20 06:57:25.705297823 +0000 UTC m=+1002.476586801" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.709999 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd72cda4-24aa-4f03-ba43-767587bb7925","Type":"ContainerStarted","Data":"31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.719369 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.720028 4492 generic.go:334] "Generic (PLEG): container finished" podID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerID="a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18" exitCode=0 Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.720052 4492 generic.go:334] "Generic (PLEG): container finished" podID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" containerID="eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a" exitCode=143 Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.720112 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"504bb2ee-6c5c-4858-ba94-88ccc1267e5d","Type":"ContainerDied","Data":"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.720142 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"504bb2ee-6c5c-4858-ba94-88ccc1267e5d","Type":"ContainerDied","Data":"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.720153 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"504bb2ee-6c5c-4858-ba94-88ccc1267e5d","Type":"ContainerDied","Data":"de4315f67f7a989ea29398a40cb76f302696a4d57bcf8ffdb1aae92a6c7f57c1"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.720173 4492 scope.go:117] "RemoveContainer" containerID="a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.720356 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.735238 4492 generic.go:334] "Generic (PLEG): container finished" podID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerID="28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f" exitCode=0 Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.735428 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerDied","Data":"28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.735496 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0","Type":"ContainerDied","Data":"6623062694fe630efe7592edb65f3cd7073b983a25176c0eb4c520c7cf601154"} Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.736279 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.741983 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-f5cc678f7-42jkm" podStartSLOduration=3.881699089 podStartE2EDuration="7.741972653s" podCreationTimestamp="2026-02-20 06:57:18 +0000 UTC" firstStartedPulling="2026-02-20 06:57:20.464506653 +0000 UTC m=+997.235795630" lastFinishedPulling="2026-02-20 06:57:24.324780216 +0000 UTC m=+1001.096069194" observedRunningTime="2026-02-20 06:57:25.729150617 +0000 UTC m=+1002.500439596" watchObservedRunningTime="2026-02-20 06:57:25.741972653 +0000 UTC m=+1002.513261631" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.764916 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.802457 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.812599 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.816736 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.817398 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="ceilometer-notification-agent" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.817418 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="ceilometer-notification-agent" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.817462 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="proxy-httpd" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.817469 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="proxy-httpd" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.817541 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="sg-core" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.817549 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="sg-core" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.817940 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="sg-core" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.817961 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="proxy-httpd" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.817989 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" containerName="ceilometer-notification-agent" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.819356 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.820628 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.821750 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.821903 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.828050 4492 scope.go:117] "RemoveContainer" containerID="eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.828856 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.610737662 podStartE2EDuration="6.828847608s" podCreationTimestamp="2026-02-20 06:57:19 +0000 UTC" firstStartedPulling="2026-02-20 06:57:20.963349604 +0000 UTC m=+997.734638582" lastFinishedPulling="2026-02-20 06:57:22.18145955 +0000 UTC m=+998.952748528" observedRunningTime="2026-02-20 06:57:25.78587654 +0000 UTC m=+1002.557165508" watchObservedRunningTime="2026-02-20 06:57:25.828847608 +0000 UTC m=+1002.600136575" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.831459 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-log-httpd\") pod \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.831537 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-run-httpd\") pod \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.831860 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" (UID: "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.832012 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-sg-core-conf-yaml\") pod \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.832102 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-combined-ca-bundle\") pod \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.832126 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-scripts\") pod \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.832152 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb8wp\" (UniqueName: \"kubernetes.io/projected/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-kube-api-access-gb8wp\") pod \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.832206 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-config-data\") pod \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\" (UID: \"0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0\") " Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.832870 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.833641 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" (UID: "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.856452 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-scripts" (OuterVolumeSpecName: "scripts") pod "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" (UID: "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.856558 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.876955 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-kube-api-access-gb8wp" (OuterVolumeSpecName: "kube-api-access-gb8wp") pod "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" (UID: "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0"). InnerVolumeSpecName "kube-api-access-gb8wp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.896315 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" (UID: "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.905158 4492 scope.go:117] "RemoveContainer" containerID="a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.906261 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18\": container with ID starting with a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18 not found: ID does not exist" containerID="a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.906296 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18"} err="failed to get container status \"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18\": rpc error: code = NotFound desc = could not find container \"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18\": container with ID starting with a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18 not found: ID does not exist" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.906317 4492 scope.go:117] "RemoveContainer" containerID="eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a" Feb 20 06:57:25 crc kubenswrapper[4492]: E0220 06:57:25.906809 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a\": container with ID starting with eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a not found: ID does not exist" containerID="eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.906836 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a"} err="failed to get container status \"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a\": rpc error: code = NotFound desc = could not find container \"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a\": container with ID starting with eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a not found: ID does not exist" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.906852 4492 scope.go:117] "RemoveContainer" containerID="a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.907333 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18"} err="failed to get container status \"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18\": rpc error: code = NotFound desc = could not find container \"a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18\": container with ID starting with a8abe600e7ac10cc05c2641d9f1e0438ee475daef0867e4d16d62a1a9f462a18 not found: ID does not exist" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.907350 4492 scope.go:117] "RemoveContainer" containerID="eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.908070 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a"} err="failed to get container status \"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a\": rpc error: code = NotFound desc = could not find container \"eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a\": container with ID starting with eefac56c1cd2a9bc8f02f8cb5ef807ca2844a33817bedc211bfd3617186e9d0a not found: ID does not exist" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.908093 4492 scope.go:117] "RemoveContainer" containerID="64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.917652 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" (UID: "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936160 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936230 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936275 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-config-data-custom\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936298 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpnq6\" (UniqueName: \"kubernetes.io/projected/86f25515-0393-402b-a3f9-1c787e7e911e-kube-api-access-xpnq6\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936338 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f25515-0393-402b-a3f9-1c787e7e911e-logs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936359 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86f25515-0393-402b-a3f9-1c787e7e911e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936389 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936413 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-config-data\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936448 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-scripts\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936514 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936531 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936543 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936553 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936563 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb8wp\" (UniqueName: \"kubernetes.io/projected/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-kube-api-access-gb8wp\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.936639 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-config-data" (OuterVolumeSpecName: "config-data") pod "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" (UID: "0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.946457 4492 scope.go:117] "RemoveContainer" containerID="516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032" Feb 20 06:57:25 crc kubenswrapper[4492]: I0220 06:57:25.988215 4492 scope.go:117] "RemoveContainer" containerID="28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.041762 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042079 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-config-data-custom\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042111 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpnq6\" (UniqueName: \"kubernetes.io/projected/86f25515-0393-402b-a3f9-1c787e7e911e-kube-api-access-xpnq6\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042178 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f25515-0393-402b-a3f9-1c787e7e911e-logs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042203 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86f25515-0393-402b-a3f9-1c787e7e911e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042279 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042303 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-config-data\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042367 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-scripts\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042417 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.042515 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.045807 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86f25515-0393-402b-a3f9-1c787e7e911e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.046343 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f25515-0393-402b-a3f9-1c787e7e911e-logs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.041911 4492 scope.go:117] "RemoveContainer" containerID="64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.048441 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: E0220 06:57:26.052929 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb\": container with ID starting with 64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb not found: ID does not exist" containerID="64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.052960 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb"} err="failed to get container status \"64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb\": rpc error: code = NotFound desc = could not find container \"64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb\": container with ID starting with 64bf74b763e0bf62f653922895fd2b2dd29529f21837a6f9bef0318118deeacb not found: ID does not exist" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.052994 4492 scope.go:117] "RemoveContainer" containerID="516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.053781 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-config-data\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: E0220 06:57:26.053854 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032\": container with ID starting with 516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032 not found: ID does not exist" containerID="516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.053874 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032"} err="failed to get container status \"516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032\": rpc error: code = NotFound desc = could not find container \"516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032\": container with ID starting with 516788e82716dbfae354ed5c8f51282f2e00ffad8892444585241a257fe9d032 not found: ID does not exist" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.053889 4492 scope.go:117] "RemoveContainer" containerID="28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.054194 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-scripts\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: E0220 06:57:26.058967 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f\": container with ID starting with 28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f not found: ID does not exist" containerID="28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.058986 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f"} err="failed to get container status \"28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f\": rpc error: code = NotFound desc = could not find container \"28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f\": container with ID starting with 28b87689bd8efe2fb54f18a8b17ce066adb1ce5d95fda21260c43b381ae9691f not found: ID does not exist" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.063402 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.065972 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-config-data-custom\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.066944 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f25515-0393-402b-a3f9-1c787e7e911e-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.068104 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpnq6\" (UniqueName: \"kubernetes.io/projected/86f25515-0393-402b-a3f9-1c787e7e911e-kube-api-access-xpnq6\") pod \"cinder-api-0\" (UID: \"86f25515-0393-402b-a3f9-1c787e7e911e\") " pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.143325 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.156395 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.156904 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.165533 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.167632 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.169793 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.170890 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.189164 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.246221 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czj99\" (UniqueName: \"kubernetes.io/projected/0318c0cb-0268-4818-872b-b9a194dc0622-kube-api-access-czj99\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.246281 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.246309 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.246335 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-log-httpd\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.246353 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-run-httpd\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.246378 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-config-data\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.246426 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-scripts\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.348215 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czj99\" (UniqueName: \"kubernetes.io/projected/0318c0cb-0268-4818-872b-b9a194dc0622-kube-api-access-czj99\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.348441 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.348488 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.348520 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-log-httpd\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.348536 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-run-httpd\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.348564 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-config-data\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.348626 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-scripts\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.349100 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-log-httpd\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.349864 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-run-httpd\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.356499 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-config-data\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.358023 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.358103 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.358955 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-scripts\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.388180 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czj99\" (UniqueName: \"kubernetes.io/projected/0318c0cb-0268-4818-872b-b9a194dc0622-kube-api-access-czj99\") pod \"ceilometer-0\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.423909 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9db58c4f4-5prd7"] Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.492146 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.693383 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 20 06:57:26 crc kubenswrapper[4492]: W0220 06:57:26.695029 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86f25515_0393_402b_a3f9_1c787e7e911e.slice/crio-d2ef6b4dbacbc164f87bc830e21af2759ee1d066ed9c984971a76b9be8d4627c WatchSource:0}: Error finding container d2ef6b4dbacbc164f87bc830e21af2759ee1d066ed9c984971a76b9be8d4627c: Status 404 returned error can't find the container with id d2ef6b4dbacbc164f87bc830e21af2759ee1d066ed9c984971a76b9be8d4627c Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.771550 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9db58c4f4-5prd7" event={"ID":"4e3149f8-a4fe-484c-a0eb-9799994a144c","Type":"ContainerStarted","Data":"32c0c835b9f2e2e143b0d169712c06d4d1dec02076952c399d0a6159c2d38e8e"} Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.771601 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9db58c4f4-5prd7" event={"ID":"4e3149f8-a4fe-484c-a0eb-9799994a144c","Type":"ContainerStarted","Data":"ea08785b678dda2c78c6423efb8b95b41bd7e957116a18502dc37927582fabcf"} Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.772768 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86f25515-0393-402b-a3f9-1c787e7e911e","Type":"ContainerStarted","Data":"d2ef6b4dbacbc164f87bc830e21af2759ee1d066ed9c984971a76b9be8d4627c"} Feb 20 06:57:26 crc kubenswrapper[4492]: I0220 06:57:26.969543 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.571611 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0" path="/var/lib/kubelet/pods/0bc02bf3-f6cd-4d60-ade4-2b19f7f9cff0/volumes" Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.572601 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="504bb2ee-6c5c-4858-ba94-88ccc1267e5d" path="/var/lib/kubelet/pods/504bb2ee-6c5c-4858-ba94-88ccc1267e5d/volumes" Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.788938 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9db58c4f4-5prd7" event={"ID":"4e3149f8-a4fe-484c-a0eb-9799994a144c","Type":"ContainerStarted","Data":"7a9c71dc3da60c897e5b19b270a58abe79cb1da2c7c43d7685e90904ee4d0bda"} Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.790637 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.790843 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.792916 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86f25515-0393-402b-a3f9-1c787e7e911e","Type":"ContainerStarted","Data":"823f96653db52a754a25ef6ee0ae66ccc21106ec3e0d65fb0df48b53a0af55a3"} Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.793722 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerStarted","Data":"eb2b836f739684704a9ce37879b11b264ca980a8ce63974dfd1df281a2bbd703"} Feb 20 06:57:27 crc kubenswrapper[4492]: I0220 06:57:27.817527 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9db58c4f4-5prd7" podStartSLOduration=2.817511078 podStartE2EDuration="2.817511078s" podCreationTimestamp="2026-02-20 06:57:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:27.810351926 +0000 UTC m=+1004.581640904" watchObservedRunningTime="2026-02-20 06:57:27.817511078 +0000 UTC m=+1004.588800057" Feb 20 06:57:28 crc kubenswrapper[4492]: I0220 06:57:28.507244 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7pcm9" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="registry-server" probeResult="failure" output=< Feb 20 06:57:28 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 06:57:28 crc kubenswrapper[4492]: > Feb 20 06:57:28 crc kubenswrapper[4492]: I0220 06:57:28.802132 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerStarted","Data":"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704"} Feb 20 06:57:28 crc kubenswrapper[4492]: I0220 06:57:28.802179 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerStarted","Data":"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229"} Feb 20 06:57:28 crc kubenswrapper[4492]: I0220 06:57:28.803900 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86f25515-0393-402b-a3f9-1c787e7e911e","Type":"ContainerStarted","Data":"79dada84d0c246827e27bbf1ceb9497c4fbf2aee58cd9cdf4e526ec634254c96"} Feb 20 06:57:28 crc kubenswrapper[4492]: I0220 06:57:28.826239 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.82622169 podStartE2EDuration="3.82622169s" podCreationTimestamp="2026-02-20 06:57:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:28.819532984 +0000 UTC m=+1005.590821962" watchObservedRunningTime="2026-02-20 06:57:28.82622169 +0000 UTC m=+1005.597510667" Feb 20 06:57:29 crc kubenswrapper[4492]: I0220 06:57:29.836569 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerStarted","Data":"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b"} Feb 20 06:57:29 crc kubenswrapper[4492]: I0220 06:57:29.836905 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.082247 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.153257 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.220949 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75dfbfdb9-5flr6"] Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.221329 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" podUID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerName="dnsmasq-dns" containerID="cri-o://eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac" gracePeriod=10 Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.304137 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.513914 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.657525 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.744756 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.757078 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-config\") pod \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.757225 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-swift-storage-0\") pod \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.757284 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-nb\") pod \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.757307 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdlvm\" (UniqueName: \"kubernetes.io/projected/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-kube-api-access-rdlvm\") pod \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.757401 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-svc\") pod \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.757467 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-sb\") pod \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\" (UID: \"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4\") " Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.781864 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-kube-api-access-rdlvm" (OuterVolumeSpecName: "kube-api-access-rdlvm") pod "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" (UID: "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4"). InnerVolumeSpecName "kube-api-access-rdlvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.843886 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" (UID: "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.852711 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" (UID: "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.858232 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" (UID: "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.860275 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.860313 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdlvm\" (UniqueName: \"kubernetes.io/projected/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-kube-api-access-rdlvm\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.860327 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.860336 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.864246 4492 generic.go:334] "Generic (PLEG): container finished" podID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerID="eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac" exitCode=0 Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.864392 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.865082 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" event={"ID":"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4","Type":"ContainerDied","Data":"eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac"} Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.865122 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75dfbfdb9-5flr6" event={"ID":"9fdb3274-7dd8-46f2-a04b-0398fc42c7a4","Type":"ContainerDied","Data":"8b13377be4dd9894076f9c8340425075f2f433d38dcd926fcdbaa484279cf923"} Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.865145 4492 scope.go:117] "RemoveContainer" containerID="eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.883696 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" (UID: "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.884599 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-config" (OuterVolumeSpecName: "config") pod "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" (UID: "9fdb3274-7dd8-46f2-a04b-0398fc42c7a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.887099 4492 scope.go:117] "RemoveContainer" containerID="137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.942773 4492 scope.go:117] "RemoveContainer" containerID="eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac" Feb 20 06:57:30 crc kubenswrapper[4492]: E0220 06:57:30.944353 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac\": container with ID starting with eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac not found: ID does not exist" containerID="eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.944388 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac"} err="failed to get container status \"eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac\": rpc error: code = NotFound desc = could not find container \"eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac\": container with ID starting with eecfbdb0e54dbe8d2bef774e7581d997d47634edeba09d77cf99dcbe02b18fac not found: ID does not exist" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.944412 4492 scope.go:117] "RemoveContainer" containerID="137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983" Feb 20 06:57:30 crc kubenswrapper[4492]: E0220 06:57:30.944666 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983\": container with ID starting with 137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983 not found: ID does not exist" containerID="137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.944688 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983"} err="failed to get container status \"137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983\": rpc error: code = NotFound desc = could not find container \"137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983\": container with ID starting with 137c289dcf908f5f7cc126272ea6ec78d4f4ba518160ba738074261924ae6983 not found: ID does not exist" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.944866 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.962298 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:30 crc kubenswrapper[4492]: I0220 06:57:30.962323 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:31 crc kubenswrapper[4492]: I0220 06:57:31.198987 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75dfbfdb9-5flr6"] Feb 20 06:57:31 crc kubenswrapper[4492]: I0220 06:57:31.211936 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75dfbfdb9-5flr6"] Feb 20 06:57:31 crc kubenswrapper[4492]: I0220 06:57:31.568785 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" path="/var/lib/kubelet/pods/9fdb3274-7dd8-46f2-a04b-0398fc42c7a4/volumes" Feb 20 06:57:31 crc kubenswrapper[4492]: I0220 06:57:31.878316 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerStarted","Data":"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b"} Feb 20 06:57:31 crc kubenswrapper[4492]: I0220 06:57:31.878585 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="cinder-scheduler" containerID="cri-o://df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018" gracePeriod=30 Feb 20 06:57:31 crc kubenswrapper[4492]: I0220 06:57:31.878638 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="probe" containerID="cri-o://31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0" gracePeriod=30 Feb 20 06:57:31 crc kubenswrapper[4492]: I0220 06:57:31.908329 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9462764030000002 podStartE2EDuration="5.908307698s" podCreationTimestamp="2026-02-20 06:57:26 +0000 UTC" firstStartedPulling="2026-02-20 06:57:26.98816147 +0000 UTC m=+1003.759450449" lastFinishedPulling="2026-02-20 06:57:30.950192766 +0000 UTC m=+1007.721481744" observedRunningTime="2026-02-20 06:57:31.90493418 +0000 UTC m=+1008.676223159" watchObservedRunningTime="2026-02-20 06:57:31.908307698 +0000 UTC m=+1008.679596677" Feb 20 06:57:32 crc kubenswrapper[4492]: I0220 06:57:32.890605 4492 generic.go:334] "Generic (PLEG): container finished" podID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerID="31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0" exitCode=0 Feb 20 06:57:32 crc kubenswrapper[4492]: I0220 06:57:32.890690 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd72cda4-24aa-4f03-ba43-767587bb7925","Type":"ContainerDied","Data":"31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0"} Feb 20 06:57:32 crc kubenswrapper[4492]: I0220 06:57:32.891219 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 06:57:33 crc kubenswrapper[4492]: I0220 06:57:33.951187 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:57:33 crc kubenswrapper[4492]: I0220 06:57:33.966049 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:57:35 crc kubenswrapper[4492]: I0220 06:57:35.469639 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:57:35 crc kubenswrapper[4492]: I0220 06:57:35.497597 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d96664ccd-bltcr" Feb 20 06:57:35 crc kubenswrapper[4492]: I0220 06:57:35.602840 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65f5db898d-bxntp"] Feb 20 06:57:35 crc kubenswrapper[4492]: I0220 06:57:35.932152 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65f5db898d-bxntp" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon-log" containerID="cri-o://953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba" gracePeriod=30 Feb 20 06:57:35 crc kubenswrapper[4492]: I0220 06:57:35.932286 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-65f5db898d-bxntp" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" containerID="cri-o://dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec" gracePeriod=30 Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.545457 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.687809 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd72cda4-24aa-4f03-ba43-767587bb7925-etc-machine-id\") pod \"bd72cda4-24aa-4f03-ba43-767587bb7925\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.687925 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjfkt\" (UniqueName: \"kubernetes.io/projected/bd72cda4-24aa-4f03-ba43-767587bb7925-kube-api-access-rjfkt\") pod \"bd72cda4-24aa-4f03-ba43-767587bb7925\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.687925 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd72cda4-24aa-4f03-ba43-767587bb7925-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bd72cda4-24aa-4f03-ba43-767587bb7925" (UID: "bd72cda4-24aa-4f03-ba43-767587bb7925"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.688094 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data-custom\") pod \"bd72cda4-24aa-4f03-ba43-767587bb7925\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.688344 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-scripts\") pod \"bd72cda4-24aa-4f03-ba43-767587bb7925\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.688421 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-combined-ca-bundle\") pod \"bd72cda4-24aa-4f03-ba43-767587bb7925\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.688536 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data\") pod \"bd72cda4-24aa-4f03-ba43-767587bb7925\" (UID: \"bd72cda4-24aa-4f03-ba43-767587bb7925\") " Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.689540 4492 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bd72cda4-24aa-4f03-ba43-767587bb7925-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.698741 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-scripts" (OuterVolumeSpecName: "scripts") pod "bd72cda4-24aa-4f03-ba43-767587bb7925" (UID: "bd72cda4-24aa-4f03-ba43-767587bb7925"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.699460 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd72cda4-24aa-4f03-ba43-767587bb7925-kube-api-access-rjfkt" (OuterVolumeSpecName: "kube-api-access-rjfkt") pod "bd72cda4-24aa-4f03-ba43-767587bb7925" (UID: "bd72cda4-24aa-4f03-ba43-767587bb7925"). InnerVolumeSpecName "kube-api-access-rjfkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.702625 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bd72cda4-24aa-4f03-ba43-767587bb7925" (UID: "bd72cda4-24aa-4f03-ba43-767587bb7925"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.769610 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd72cda4-24aa-4f03-ba43-767587bb7925" (UID: "bd72cda4-24aa-4f03-ba43-767587bb7925"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.791546 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.791576 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.791589 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjfkt\" (UniqueName: \"kubernetes.io/projected/bd72cda4-24aa-4f03-ba43-767587bb7925-kube-api-access-rjfkt\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.791601 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.823607 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data" (OuterVolumeSpecName: "config-data") pod "bd72cda4-24aa-4f03-ba43-767587bb7925" (UID: "bd72cda4-24aa-4f03-ba43-767587bb7925"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.894191 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd72cda4-24aa-4f03-ba43-767587bb7925-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.942648 4492 generic.go:334] "Generic (PLEG): container finished" podID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerID="df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018" exitCode=0 Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.942710 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd72cda4-24aa-4f03-ba43-767587bb7925","Type":"ContainerDied","Data":"df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018"} Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.942741 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bd72cda4-24aa-4f03-ba43-767587bb7925","Type":"ContainerDied","Data":"aa427d55678ecbce2f55d192d58b3291f9fea1a3ad7600f566944170a9c2bc68"} Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.942763 4492 scope.go:117] "RemoveContainer" containerID="31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.942767 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.962043 4492 scope.go:117] "RemoveContainer" containerID="df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.982624 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.990115 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.998976 4492 scope.go:117] "RemoveContainer" containerID="31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0" Feb 20 06:57:36 crc kubenswrapper[4492]: E0220 06:57:36.999537 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0\": container with ID starting with 31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0 not found: ID does not exist" containerID="31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.999701 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0"} err="failed to get container status \"31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0\": rpc error: code = NotFound desc = could not find container \"31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0\": container with ID starting with 31f3a848f3352ad2631ec649316b73588f4bfe699916048845cf4a38b2b035b0 not found: ID does not exist" Feb 20 06:57:36 crc kubenswrapper[4492]: I0220 06:57:36.999788 4492 scope.go:117] "RemoveContainer" containerID="df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018" Feb 20 06:57:37 crc kubenswrapper[4492]: E0220 06:57:37.000260 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018\": container with ID starting with df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018 not found: ID does not exist" containerID="df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.000293 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018"} err="failed to get container status \"df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018\": rpc error: code = NotFound desc = could not find container \"df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018\": container with ID starting with df367ba4a311c751d6235cc2e59f4ac1f9ec56e8f0e75e84b3f9349396f75018 not found: ID does not exist" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.011059 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:37 crc kubenswrapper[4492]: E0220 06:57:37.011491 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="probe" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.011573 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="probe" Feb 20 06:57:37 crc kubenswrapper[4492]: E0220 06:57:37.011641 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerName="dnsmasq-dns" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.011705 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerName="dnsmasq-dns" Feb 20 06:57:37 crc kubenswrapper[4492]: E0220 06:57:37.011767 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="cinder-scheduler" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.011833 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="cinder-scheduler" Feb 20 06:57:37 crc kubenswrapper[4492]: E0220 06:57:37.011891 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerName="init" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.011940 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerName="init" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.012146 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdb3274-7dd8-46f2-a04b-0398fc42c7a4" containerName="dnsmasq-dns" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.015594 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="probe" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.015716 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" containerName="cinder-scheduler" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.016750 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.019592 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.033343 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.199405 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.199465 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.199509 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dprlq\" (UniqueName: \"kubernetes.io/projected/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-kube-api-access-dprlq\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.199541 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.199580 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.199600 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.301017 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.301077 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.301179 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.301222 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.301250 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dprlq\" (UniqueName: \"kubernetes.io/projected/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-kube-api-access-dprlq\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.301288 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.301386 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.307989 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.310210 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.315334 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.322907 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.337787 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dprlq\" (UniqueName: \"kubernetes.io/projected/343b4ca3-ca66-42d1-9b10-8c5d5c9863e5-kube-api-access-dprlq\") pod \"cinder-scheduler-0\" (UID: \"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5\") " pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.450315 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.516272 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.565639 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd72cda4-24aa-4f03-ba43-767587bb7925" path="/var/lib/kubelet/pods/bd72cda4-24aa-4f03-ba43-767587bb7925/volumes" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.566444 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.632062 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.801211 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7pcm9"] Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.923972 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 20 06:57:37 crc kubenswrapper[4492]: I0220 06:57:37.987779 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5","Type":"ContainerStarted","Data":"306da636bccb4236aaa18c880f19a083406972cd09248e2dcf3121a6021d1b7f"} Feb 20 06:57:38 crc kubenswrapper[4492]: I0220 06:57:38.474519 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9db58c4f4-5prd7" Feb 20 06:57:38 crc kubenswrapper[4492]: I0220 06:57:38.540687 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7547fb57c8-ndpch"] Feb 20 06:57:38 crc kubenswrapper[4492]: I0220 06:57:38.541807 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7547fb57c8-ndpch" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api" containerID="cri-o://b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c" gracePeriod=30 Feb 20 06:57:38 crc kubenswrapper[4492]: I0220 06:57:38.541886 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7547fb57c8-ndpch" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api-log" containerID="cri-o://f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4" gracePeriod=30 Feb 20 06:57:38 crc kubenswrapper[4492]: I0220 06:57:38.956131 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:38 crc kubenswrapper[4492]: I0220 06:57:38.972834 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.026725 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5","Type":"ContainerStarted","Data":"4c908089794424f17fb9607f27ca2ce415bd67e424424ee41ae9275dd6253b7b"} Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.032410 4492 generic.go:334] "Generic (PLEG): container finished" podID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerID="f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4" exitCode=143 Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.033407 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7547fb57c8-ndpch" event={"ID":"5de1b303-1a15-44d9-b3fd-dfc3381dc491","Type":"ContainerDied","Data":"f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4"} Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.033557 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7pcm9" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="registry-server" containerID="cri-o://50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8" gracePeriod=2 Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.412438 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.706322 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.721879 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7fcbbc74fb-jdvms" Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.892172 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7dcr\" (UniqueName: \"kubernetes.io/projected/3ed937a1-e598-4418-81f6-cbbd240e4d4a-kube-api-access-b7dcr\") pod \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.892569 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-utilities\") pod \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.892822 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-catalog-content\") pod \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\" (UID: \"3ed937a1-e598-4418-81f6-cbbd240e4d4a\") " Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.893446 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-utilities" (OuterVolumeSpecName: "utilities") pod "3ed937a1-e598-4418-81f6-cbbd240e4d4a" (UID: "3ed937a1-e598-4418-81f6-cbbd240e4d4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.895062 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:39 crc kubenswrapper[4492]: I0220 06:57:39.905541 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ed937a1-e598-4418-81f6-cbbd240e4d4a-kube-api-access-b7dcr" (OuterVolumeSpecName: "kube-api-access-b7dcr") pod "3ed937a1-e598-4418-81f6-cbbd240e4d4a" (UID: "3ed937a1-e598-4418-81f6-cbbd240e4d4a"). InnerVolumeSpecName "kube-api-access-b7dcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.004162 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7dcr\" (UniqueName: \"kubernetes.io/projected/3ed937a1-e598-4418-81f6-cbbd240e4d4a-kube-api-access-b7dcr\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.020723 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ed937a1-e598-4418-81f6-cbbd240e4d4a" (UID: "3ed937a1-e598-4418-81f6-cbbd240e4d4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.067449 4492 generic.go:334] "Generic (PLEG): container finished" podID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerID="50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8" exitCode=0 Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.067538 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pcm9" event={"ID":"3ed937a1-e598-4418-81f6-cbbd240e4d4a","Type":"ContainerDied","Data":"50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8"} Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.067589 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7pcm9" event={"ID":"3ed937a1-e598-4418-81f6-cbbd240e4d4a","Type":"ContainerDied","Data":"fb77c7cac457fee555a8f11ca3fe822d8842192ce7cf9c521bcfb14adee3afc1"} Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.067613 4492 scope.go:117] "RemoveContainer" containerID="50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.067784 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7pcm9" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.086231 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"343b4ca3-ca66-42d1-9b10-8c5d5c9863e5","Type":"ContainerStarted","Data":"1b29466fe3b3fadc7eb4a9f19752d5ccb07a2735c019991f6f0ceebf5f580c53"} Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.105516 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ed937a1-e598-4418-81f6-cbbd240e4d4a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.109200 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7pcm9"] Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.110368 4492 generic.go:334] "Generic (PLEG): container finished" podID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerID="dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec" exitCode=0 Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.110417 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65f5db898d-bxntp" event={"ID":"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7","Type":"ContainerDied","Data":"dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec"} Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.123586 4492 scope.go:117] "RemoveContainer" containerID="8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.129022 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7pcm9"] Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.131368 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.131359248 podStartE2EDuration="4.131359248s" podCreationTimestamp="2026-02-20 06:57:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:40.130879563 +0000 UTC m=+1016.902168542" watchObservedRunningTime="2026-02-20 06:57:40.131359248 +0000 UTC m=+1016.902648225" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.151610 4492 scope.go:117] "RemoveContainer" containerID="9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.172764 4492 scope.go:117] "RemoveContainer" containerID="50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8" Feb 20 06:57:40 crc kubenswrapper[4492]: E0220 06:57:40.173147 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8\": container with ID starting with 50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8 not found: ID does not exist" containerID="50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.173271 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8"} err="failed to get container status \"50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8\": rpc error: code = NotFound desc = could not find container \"50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8\": container with ID starting with 50ecea3eaaa50945ad96698bb947f7e34d46c449da04c815070803a2f6bbaaf8 not found: ID does not exist" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.173365 4492 scope.go:117] "RemoveContainer" containerID="8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d" Feb 20 06:57:40 crc kubenswrapper[4492]: E0220 06:57:40.173842 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d\": container with ID starting with 8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d not found: ID does not exist" containerID="8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.173884 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d"} err="failed to get container status \"8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d\": rpc error: code = NotFound desc = could not find container \"8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d\": container with ID starting with 8f95ee6d312123203d10c7930d8f404232a818cf20a6e3a4d386de817bcea69d not found: ID does not exist" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.173905 4492 scope.go:117] "RemoveContainer" containerID="9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8" Feb 20 06:57:40 crc kubenswrapper[4492]: E0220 06:57:40.174374 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8\": container with ID starting with 9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8 not found: ID does not exist" containerID="9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.174397 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8"} err="failed to get container status \"9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8\": rpc error: code = NotFound desc = could not find container \"9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8\": container with ID starting with 9364155c848d3fff3f4467e7f64dc373ef9833c609db573ee7e0c6cb5ce7add8 not found: ID does not exist" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.587267 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.605439 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65bf8dfc4b-nwxp9" Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.655705 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-565b895f9d-w6nzp"] Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.655927 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-565b895f9d-w6nzp" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-log" containerID="cri-o://78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2" gracePeriod=30 Feb 20 06:57:40 crc kubenswrapper[4492]: I0220 06:57:40.656295 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-565b895f9d-w6nzp" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-api" containerID="cri-o://7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199" gracePeriod=30 Feb 20 06:57:41 crc kubenswrapper[4492]: I0220 06:57:41.120357 4492 generic.go:334] "Generic (PLEG): container finished" podID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerID="78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2" exitCode=143 Feb 20 06:57:41 crc kubenswrapper[4492]: I0220 06:57:41.120434 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565b895f9d-w6nzp" event={"ID":"87b134e6-2d14-4fc6-bfd9-6055300dab55","Type":"ContainerDied","Data":"78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2"} Feb 20 06:57:41 crc kubenswrapper[4492]: I0220 06:57:41.266362 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65f5db898d-bxntp" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 20 06:57:41 crc kubenswrapper[4492]: I0220 06:57:41.567704 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" path="/var/lib/kubelet/pods/3ed937a1-e598-4418-81f6-cbbd240e4d4a/volumes" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.088058 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7547fb57c8-ndpch" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.167:9311/healthcheck\": read tcp 10.217.0.2:53598->10.217.0.167:9311: read: connection reset by peer" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.088083 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7547fb57c8-ndpch" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.167:9311/healthcheck\": read tcp 10.217.0.2:53602->10.217.0.167:9311: read: connection reset by peer" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.586336 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.633157 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.654348 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data-custom\") pod \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.654440 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnz45\" (UniqueName: \"kubernetes.io/projected/5de1b303-1a15-44d9-b3fd-dfc3381dc491-kube-api-access-gnz45\") pod \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.654516 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data\") pod \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.654557 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5de1b303-1a15-44d9-b3fd-dfc3381dc491-logs\") pod \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.654635 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-combined-ca-bundle\") pod \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\" (UID: \"5de1b303-1a15-44d9-b3fd-dfc3381dc491\") " Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.655067 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5de1b303-1a15-44d9-b3fd-dfc3381dc491-logs" (OuterVolumeSpecName: "logs") pod "5de1b303-1a15-44d9-b3fd-dfc3381dc491" (UID: "5de1b303-1a15-44d9-b3fd-dfc3381dc491"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.662515 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de1b303-1a15-44d9-b3fd-dfc3381dc491-kube-api-access-gnz45" (OuterVolumeSpecName: "kube-api-access-gnz45") pod "5de1b303-1a15-44d9-b3fd-dfc3381dc491" (UID: "5de1b303-1a15-44d9-b3fd-dfc3381dc491"). InnerVolumeSpecName "kube-api-access-gnz45". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.680322 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5de1b303-1a15-44d9-b3fd-dfc3381dc491" (UID: "5de1b303-1a15-44d9-b3fd-dfc3381dc491"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.690421 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5de1b303-1a15-44d9-b3fd-dfc3381dc491" (UID: "5de1b303-1a15-44d9-b3fd-dfc3381dc491"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.722577 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data" (OuterVolumeSpecName: "config-data") pod "5de1b303-1a15-44d9-b3fd-dfc3381dc491" (UID: "5de1b303-1a15-44d9-b3fd-dfc3381dc491"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.757266 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.757298 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.757310 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnz45\" (UniqueName: \"kubernetes.io/projected/5de1b303-1a15-44d9-b3fd-dfc3381dc491-kube-api-access-gnz45\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.757322 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de1b303-1a15-44d9-b3fd-dfc3381dc491-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:42 crc kubenswrapper[4492]: I0220 06:57:42.757332 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5de1b303-1a15-44d9-b3fd-dfc3381dc491-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.156789 4492 generic.go:334] "Generic (PLEG): container finished" podID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerID="b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c" exitCode=0 Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.156843 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7547fb57c8-ndpch" event={"ID":"5de1b303-1a15-44d9-b3fd-dfc3381dc491","Type":"ContainerDied","Data":"b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c"} Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.156877 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7547fb57c8-ndpch" event={"ID":"5de1b303-1a15-44d9-b3fd-dfc3381dc491","Type":"ContainerDied","Data":"439bbc88010ae80f02e83a26cda2fa2dbcd75f9755d99945cc10255b89df4020"} Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.156908 4492 scope.go:117] "RemoveContainer" containerID="b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.157086 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7547fb57c8-ndpch" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.222131 4492 scope.go:117] "RemoveContainer" containerID="f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.222287 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7547fb57c8-ndpch"] Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.240007 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7547fb57c8-ndpch"] Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.269192 4492 scope.go:117] "RemoveContainer" containerID="b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c" Feb 20 06:57:43 crc kubenswrapper[4492]: E0220 06:57:43.271315 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c\": container with ID starting with b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c not found: ID does not exist" containerID="b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.271367 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c"} err="failed to get container status \"b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c\": rpc error: code = NotFound desc = could not find container \"b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c\": container with ID starting with b0c677b9a923429915bc560f7d986d1d5c6238518c50c84742a0c31ff320308c not found: ID does not exist" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.271402 4492 scope.go:117] "RemoveContainer" containerID="f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4" Feb 20 06:57:43 crc kubenswrapper[4492]: E0220 06:57:43.274146 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4\": container with ID starting with f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4 not found: ID does not exist" containerID="f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.274180 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4"} err="failed to get container status \"f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4\": rpc error: code = NotFound desc = could not find container \"f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4\": container with ID starting with f2bb522430f1eb0e33e34f830c67822f821379f5a7d0f075d7e033886f1005b4 not found: ID does not exist" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.570077 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" path="/var/lib/kubelet/pods/5de1b303-1a15-44d9-b3fd-dfc3381dc491/volumes" Feb 20 06:57:43 crc kubenswrapper[4492]: I0220 06:57:43.948981 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-65749686ff-sv9sq" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.023484 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cf7fbb848-5c8g6"] Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.023764 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cf7fbb848-5c8g6" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-api" containerID="cri-o://b351db4079da572e45b8ead2fa564d38656849c6e0ce66ea7c1fbcbdb0671376" gracePeriod=30 Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.024233 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6cf7fbb848-5c8g6" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-httpd" containerID="cri-o://f8b0f7955b6291ba776eb73ad0cea1f5f6d1f8b4ce4d244cf6a302da792fbc0f" gracePeriod=30 Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.053686 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 20 06:57:44 crc kubenswrapper[4492]: E0220 06:57:44.054114 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api-log" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054134 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api-log" Feb 20 06:57:44 crc kubenswrapper[4492]: E0220 06:57:44.054165 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054221 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api" Feb 20 06:57:44 crc kubenswrapper[4492]: E0220 06:57:44.054242 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="registry-server" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054252 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="registry-server" Feb 20 06:57:44 crc kubenswrapper[4492]: E0220 06:57:44.054287 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="extract-content" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054292 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="extract-content" Feb 20 06:57:44 crc kubenswrapper[4492]: E0220 06:57:44.054306 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="extract-utilities" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054312 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="extract-utilities" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054590 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ed937a1-e598-4418-81f6-cbbd240e4d4a" containerName="registry-server" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054629 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api-log" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.054640 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de1b303-1a15-44d9-b3fd-dfc3381dc491" containerName="barbican-api" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.055346 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.057012 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.057253 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-gtxjs" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.058720 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.064084 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.124107 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr6t8\" (UniqueName: \"kubernetes.io/projected/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-kube-api-access-cr6t8\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.124325 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-openstack-config-secret\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.124424 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.124484 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-openstack-config\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.190403 4492 generic.go:334] "Generic (PLEG): container finished" podID="3a568bef-11fc-479e-bf90-4c2030405335" containerID="f8b0f7955b6291ba776eb73ad0cea1f5f6d1f8b4ce4d244cf6a302da792fbc0f" exitCode=0 Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.190462 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf7fbb848-5c8g6" event={"ID":"3a568bef-11fc-479e-bf90-4c2030405335","Type":"ContainerDied","Data":"f8b0f7955b6291ba776eb73ad0cea1f5f6d1f8b4ce4d244cf6a302da792fbc0f"} Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.229603 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-openstack-config-secret\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.229679 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.229722 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-openstack-config\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.229913 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr6t8\" (UniqueName: \"kubernetes.io/projected/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-kube-api-access-cr6t8\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.231500 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-openstack-config\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.236586 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-openstack-config-secret\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.239025 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.247737 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr6t8\" (UniqueName: \"kubernetes.io/projected/aa29dc13-c6f5-43ef-b0a9-44b6735922d7-kube-api-access-cr6t8\") pod \"openstackclient\" (UID: \"aa29dc13-c6f5-43ef-b0a9-44b6735922d7\") " pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.376390 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.756463 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.847975 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-combined-ca-bundle\") pod \"87b134e6-2d14-4fc6-bfd9-6055300dab55\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.848099 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-internal-tls-certs\") pod \"87b134e6-2d14-4fc6-bfd9-6055300dab55\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.848167 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-config-data\") pod \"87b134e6-2d14-4fc6-bfd9-6055300dab55\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.848202 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b134e6-2d14-4fc6-bfd9-6055300dab55-logs\") pod \"87b134e6-2d14-4fc6-bfd9-6055300dab55\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.848332 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-scripts\") pod \"87b134e6-2d14-4fc6-bfd9-6055300dab55\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.848350 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-public-tls-certs\") pod \"87b134e6-2d14-4fc6-bfd9-6055300dab55\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.848489 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfzbr\" (UniqueName: \"kubernetes.io/projected/87b134e6-2d14-4fc6-bfd9-6055300dab55-kube-api-access-gfzbr\") pod \"87b134e6-2d14-4fc6-bfd9-6055300dab55\" (UID: \"87b134e6-2d14-4fc6-bfd9-6055300dab55\") " Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.851223 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87b134e6-2d14-4fc6-bfd9-6055300dab55-logs" (OuterVolumeSpecName: "logs") pod "87b134e6-2d14-4fc6-bfd9-6055300dab55" (UID: "87b134e6-2d14-4fc6-bfd9-6055300dab55"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.856820 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87b134e6-2d14-4fc6-bfd9-6055300dab55-kube-api-access-gfzbr" (OuterVolumeSpecName: "kube-api-access-gfzbr") pod "87b134e6-2d14-4fc6-bfd9-6055300dab55" (UID: "87b134e6-2d14-4fc6-bfd9-6055300dab55"). InnerVolumeSpecName "kube-api-access-gfzbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.867761 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-scripts" (OuterVolumeSpecName: "scripts") pod "87b134e6-2d14-4fc6-bfd9-6055300dab55" (UID: "87b134e6-2d14-4fc6-bfd9-6055300dab55"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.919888 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-config-data" (OuterVolumeSpecName: "config-data") pod "87b134e6-2d14-4fc6-bfd9-6055300dab55" (UID: "87b134e6-2d14-4fc6-bfd9-6055300dab55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.933553 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87b134e6-2d14-4fc6-bfd9-6055300dab55" (UID: "87b134e6-2d14-4fc6-bfd9-6055300dab55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.948829 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.951091 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.952017 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.952077 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b134e6-2d14-4fc6-bfd9-6055300dab55-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.952126 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.952179 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfzbr\" (UniqueName: \"kubernetes.io/projected/87b134e6-2d14-4fc6-bfd9-6055300dab55-kube-api-access-gfzbr\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.964829 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "87b134e6-2d14-4fc6-bfd9-6055300dab55" (UID: "87b134e6-2d14-4fc6-bfd9-6055300dab55"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:44 crc kubenswrapper[4492]: I0220 06:57:44.966585 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "87b134e6-2d14-4fc6-bfd9-6055300dab55" (UID: "87b134e6-2d14-4fc6-bfd9-6055300dab55"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.055366 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.055405 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b134e6-2d14-4fc6-bfd9-6055300dab55-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.075391 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.075959 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-central-agent" containerID="cri-o://f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" gracePeriod=30 Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.076051 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="sg-core" containerID="cri-o://9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" gracePeriod=30 Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.076201 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-notification-agent" containerID="cri-o://c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" gracePeriod=30 Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.076294 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="proxy-httpd" containerID="cri-o://171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" gracePeriod=30 Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.098912 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.173:3000/\": EOF" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.204552 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"aa29dc13-c6f5-43ef-b0a9-44b6735922d7","Type":"ContainerStarted","Data":"af0f73d5d8aa85b71504e709d6cfd434d4680d6ca3aee61d1f967d1852f024d3"} Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.207983 4492 generic.go:334] "Generic (PLEG): container finished" podID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerID="7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199" exitCode=0 Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.208032 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565b895f9d-w6nzp" event={"ID":"87b134e6-2d14-4fc6-bfd9-6055300dab55","Type":"ContainerDied","Data":"7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199"} Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.208107 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-565b895f9d-w6nzp" event={"ID":"87b134e6-2d14-4fc6-bfd9-6055300dab55","Type":"ContainerDied","Data":"1c7bd0523be39fc059a7d34142485b485e8b0a35ff121c7f30ff0143702c239b"} Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.208115 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-565b895f9d-w6nzp" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.208135 4492 scope.go:117] "RemoveContainer" containerID="7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.234751 4492 scope.go:117] "RemoveContainer" containerID="78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.294965 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-565b895f9d-w6nzp"] Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.323917 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7c8d654db7-jfjgs"] Feb 20 06:57:45 crc kubenswrapper[4492]: E0220 06:57:45.328961 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-api" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.329045 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-api" Feb 20 06:57:45 crc kubenswrapper[4492]: E0220 06:57:45.329106 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-log" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.329152 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-log" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.348806 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-log" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.348915 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" containerName="placement-api" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.351751 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.358916 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.360607 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.361103 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.380177 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-565b895f9d-w6nzp"] Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.392284 4492 scope.go:117] "RemoveContainer" containerID="7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199" Feb 20 06:57:45 crc kubenswrapper[4492]: E0220 06:57:45.394002 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199\": container with ID starting with 7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199 not found: ID does not exist" containerID="7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.394102 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199"} err="failed to get container status \"7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199\": rpc error: code = NotFound desc = could not find container \"7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199\": container with ID starting with 7c329790faf54b84736024cebbdab7a7babe69d5a1803c6ccc04acd8a4c54199 not found: ID does not exist" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.394200 4492 scope.go:117] "RemoveContainer" containerID="78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2" Feb 20 06:57:45 crc kubenswrapper[4492]: E0220 06:57:45.396346 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2\": container with ID starting with 78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2 not found: ID does not exist" containerID="78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.396424 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2"} err="failed to get container status \"78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2\": rpc error: code = NotFound desc = could not find container \"78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2\": container with ID starting with 78fcc24b6e2eae0a63c4f73dc1a2c552fe53c6e531a96891cc51320f452339d2 not found: ID does not exist" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.412519 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c8d654db7-jfjgs"] Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.478921 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-config-data\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.479001 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b14dc392-054d-4257-b340-642e7527aa67-log-httpd\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.479092 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjwzf\" (UniqueName: \"kubernetes.io/projected/b14dc392-054d-4257-b340-642e7527aa67-kube-api-access-zjwzf\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.479136 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b14dc392-054d-4257-b340-642e7527aa67-etc-swift\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.479188 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b14dc392-054d-4257-b340-642e7527aa67-run-httpd\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.479240 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-internal-tls-certs\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.479265 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-combined-ca-bundle\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.479283 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-public-tls-certs\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.567573 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87b134e6-2d14-4fc6-bfd9-6055300dab55" path="/var/lib/kubelet/pods/87b134e6-2d14-4fc6-bfd9-6055300dab55/volumes" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580377 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-config-data\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580421 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b14dc392-054d-4257-b340-642e7527aa67-log-httpd\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580497 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjwzf\" (UniqueName: \"kubernetes.io/projected/b14dc392-054d-4257-b340-642e7527aa67-kube-api-access-zjwzf\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580520 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b14dc392-054d-4257-b340-642e7527aa67-etc-swift\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580550 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b14dc392-054d-4257-b340-642e7527aa67-run-httpd\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580590 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-internal-tls-certs\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580611 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-combined-ca-bundle\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.580632 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-public-tls-certs\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.582312 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b14dc392-054d-4257-b340-642e7527aa67-run-httpd\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.582579 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b14dc392-054d-4257-b340-642e7527aa67-log-httpd\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.591643 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-internal-tls-certs\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.599367 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-config-data\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.600313 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-combined-ca-bundle\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.607582 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b14dc392-054d-4257-b340-642e7527aa67-public-tls-certs\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.608833 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b14dc392-054d-4257-b340-642e7527aa67-etc-swift\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.609180 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjwzf\" (UniqueName: \"kubernetes.io/projected/b14dc392-054d-4257-b340-642e7527aa67-kube-api-access-zjwzf\") pod \"swift-proxy-7c8d654db7-jfjgs\" (UID: \"b14dc392-054d-4257-b340-642e7527aa67\") " pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:45 crc kubenswrapper[4492]: I0220 06:57:45.703549 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.108767 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.193895 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-config-data\") pod \"0318c0cb-0268-4818-872b-b9a194dc0622\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.194022 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czj99\" (UniqueName: \"kubernetes.io/projected/0318c0cb-0268-4818-872b-b9a194dc0622-kube-api-access-czj99\") pod \"0318c0cb-0268-4818-872b-b9a194dc0622\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.194093 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-scripts\") pod \"0318c0cb-0268-4818-872b-b9a194dc0622\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.194170 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-sg-core-conf-yaml\") pod \"0318c0cb-0268-4818-872b-b9a194dc0622\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.194256 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-run-httpd\") pod \"0318c0cb-0268-4818-872b-b9a194dc0622\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.194302 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-combined-ca-bundle\") pod \"0318c0cb-0268-4818-872b-b9a194dc0622\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.194324 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-log-httpd\") pod \"0318c0cb-0268-4818-872b-b9a194dc0622\" (UID: \"0318c0cb-0268-4818-872b-b9a194dc0622\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.196017 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0318c0cb-0268-4818-872b-b9a194dc0622" (UID: "0318c0cb-0268-4818-872b-b9a194dc0622"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.196591 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0318c0cb-0268-4818-872b-b9a194dc0622" (UID: "0318c0cb-0268-4818-872b-b9a194dc0622"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.222231 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0318c0cb-0268-4818-872b-b9a194dc0622-kube-api-access-czj99" (OuterVolumeSpecName: "kube-api-access-czj99") pod "0318c0cb-0268-4818-872b-b9a194dc0622" (UID: "0318c0cb-0268-4818-872b-b9a194dc0622"). InnerVolumeSpecName "kube-api-access-czj99". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.226733 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-scripts" (OuterVolumeSpecName: "scripts") pod "0318c0cb-0268-4818-872b-b9a194dc0622" (UID: "0318c0cb-0268-4818-872b-b9a194dc0622"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.237357 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf7fbb848-5c8g6" event={"ID":"3a568bef-11fc-479e-bf90-4c2030405335","Type":"ContainerDied","Data":"b351db4079da572e45b8ead2fa564d38656849c6e0ce66ea7c1fbcbdb0671376"} Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.240748 4492 generic.go:334] "Generic (PLEG): container finished" podID="3a568bef-11fc-479e-bf90-4c2030405335" containerID="b351db4079da572e45b8ead2fa564d38656849c6e0ce66ea7c1fbcbdb0671376" exitCode=0 Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247691 4492 generic.go:334] "Generic (PLEG): container finished" podID="0318c0cb-0268-4818-872b-b9a194dc0622" containerID="171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" exitCode=0 Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247728 4492 generic.go:334] "Generic (PLEG): container finished" podID="0318c0cb-0268-4818-872b-b9a194dc0622" containerID="9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" exitCode=2 Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247737 4492 generic.go:334] "Generic (PLEG): container finished" podID="0318c0cb-0268-4818-872b-b9a194dc0622" containerID="c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" exitCode=0 Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247743 4492 generic.go:334] "Generic (PLEG): container finished" podID="0318c0cb-0268-4818-872b-b9a194dc0622" containerID="f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" exitCode=0 Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247764 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerDied","Data":"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b"} Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247794 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerDied","Data":"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b"} Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247806 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerDied","Data":"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704"} Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247815 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerDied","Data":"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229"} Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247823 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0318c0cb-0268-4818-872b-b9a194dc0622","Type":"ContainerDied","Data":"eb2b836f739684704a9ce37879b11b264ca980a8ce63974dfd1df281a2bbd703"} Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247837 4492 scope.go:117] "RemoveContainer" containerID="171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.247957 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.278215 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0318c0cb-0268-4818-872b-b9a194dc0622" (UID: "0318c0cb-0268-4818-872b-b9a194dc0622"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.296796 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czj99\" (UniqueName: \"kubernetes.io/projected/0318c0cb-0268-4818-872b-b9a194dc0622-kube-api-access-czj99\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.297038 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.297103 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.297155 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.297203 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0318c0cb-0268-4818-872b-b9a194dc0622-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.351196 4492 scope.go:117] "RemoveContainer" containerID="9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.354705 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-config-data" (OuterVolumeSpecName: "config-data") pod "0318c0cb-0268-4818-872b-b9a194dc0622" (UID: "0318c0cb-0268-4818-872b-b9a194dc0622"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.380894 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0318c0cb-0268-4818-872b-b9a194dc0622" (UID: "0318c0cb-0268-4818-872b-b9a194dc0622"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.399239 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.399267 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0318c0cb-0268-4818-872b-b9a194dc0622-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.419493 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c8d654db7-jfjgs"] Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.441289 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.478186 4492 scope.go:117] "RemoveContainer" containerID="c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.500838 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-config\") pod \"3a568bef-11fc-479e-bf90-4c2030405335\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.500914 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-ovndb-tls-certs\") pod \"3a568bef-11fc-479e-bf90-4c2030405335\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.501011 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-httpd-config\") pod \"3a568bef-11fc-479e-bf90-4c2030405335\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.501035 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26nst\" (UniqueName: \"kubernetes.io/projected/3a568bef-11fc-479e-bf90-4c2030405335-kube-api-access-26nst\") pod \"3a568bef-11fc-479e-bf90-4c2030405335\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.501119 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-combined-ca-bundle\") pod \"3a568bef-11fc-479e-bf90-4c2030405335\" (UID: \"3a568bef-11fc-479e-bf90-4c2030405335\") " Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.509165 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "3a568bef-11fc-479e-bf90-4c2030405335" (UID: "3a568bef-11fc-479e-bf90-4c2030405335"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.510558 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a568bef-11fc-479e-bf90-4c2030405335-kube-api-access-26nst" (OuterVolumeSpecName: "kube-api-access-26nst") pod "3a568bef-11fc-479e-bf90-4c2030405335" (UID: "3a568bef-11fc-479e-bf90-4c2030405335"). InnerVolumeSpecName "kube-api-access-26nst". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.551676 4492 scope.go:117] "RemoveContainer" containerID="f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.557646 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-config" (OuterVolumeSpecName: "config") pod "3a568bef-11fc-479e-bf90-4c2030405335" (UID: "3a568bef-11fc-479e-bf90-4c2030405335"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.594806 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.606404 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.606436 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.606446 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26nst\" (UniqueName: \"kubernetes.io/projected/3a568bef-11fc-479e-bf90-4c2030405335-kube-api-access-26nst\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.630573 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.635844 4492 scope.go:117] "RemoveContainer" containerID="171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.638858 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a568bef-11fc-479e-bf90-4c2030405335" (UID: "3a568bef-11fc-479e-bf90-4c2030405335"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.641134 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.642339 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-api" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642366 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-api" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.642382 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="sg-core" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642389 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="sg-core" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.642410 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-central-agent" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642417 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-central-agent" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.642428 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="proxy-httpd" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642433 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="proxy-httpd" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.642446 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-notification-agent" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642452 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-notification-agent" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.642469 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-httpd" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642490 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-httpd" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642760 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-notification-agent" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642771 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="ceilometer-central-agent" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642780 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-api" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642795 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a568bef-11fc-479e-bf90-4c2030405335" containerName="neutron-httpd" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642810 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="sg-core" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.642820 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" containerName="proxy-httpd" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.643111 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": container with ID starting with 171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b not found: ID does not exist" containerID="171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.643146 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b"} err="failed to get container status \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": rpc error: code = NotFound desc = could not find container \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": container with ID starting with 171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.643174 4492 scope.go:117] "RemoveContainer" containerID="9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.643846 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": container with ID starting with 9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b not found: ID does not exist" containerID="9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.643871 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b"} err="failed to get container status \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": rpc error: code = NotFound desc = could not find container \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": container with ID starting with 9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.643884 4492 scope.go:117] "RemoveContainer" containerID="c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.644599 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.649094 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.649296 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.654074 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": container with ID starting with c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704 not found: ID does not exist" containerID="c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654101 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704"} err="failed to get container status \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": rpc error: code = NotFound desc = could not find container \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": container with ID starting with c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654121 4492 scope.go:117] "RemoveContainer" containerID="f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" Feb 20 06:57:46 crc kubenswrapper[4492]: E0220 06:57:46.654399 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": container with ID starting with f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229 not found: ID does not exist" containerID="f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654418 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229"} err="failed to get container status \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": rpc error: code = NotFound desc = could not find container \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": container with ID starting with f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654436 4492 scope.go:117] "RemoveContainer" containerID="171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654623 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b"} err="failed to get container status \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": rpc error: code = NotFound desc = could not find container \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": container with ID starting with 171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654640 4492 scope.go:117] "RemoveContainer" containerID="9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654814 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b"} err="failed to get container status \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": rpc error: code = NotFound desc = could not find container \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": container with ID starting with 9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654828 4492 scope.go:117] "RemoveContainer" containerID="c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654977 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704"} err="failed to get container status \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": rpc error: code = NotFound desc = could not find container \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": container with ID starting with c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.654990 4492 scope.go:117] "RemoveContainer" containerID="f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655140 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229"} err="failed to get container status \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": rpc error: code = NotFound desc = could not find container \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": container with ID starting with f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655154 4492 scope.go:117] "RemoveContainer" containerID="171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655315 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b"} err="failed to get container status \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": rpc error: code = NotFound desc = could not find container \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": container with ID starting with 171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655330 4492 scope.go:117] "RemoveContainer" containerID="9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655503 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b"} err="failed to get container status \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": rpc error: code = NotFound desc = could not find container \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": container with ID starting with 9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655517 4492 scope.go:117] "RemoveContainer" containerID="c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655686 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704"} err="failed to get container status \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": rpc error: code = NotFound desc = could not find container \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": container with ID starting with c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655700 4492 scope.go:117] "RemoveContainer" containerID="f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655898 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229"} err="failed to get container status \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": rpc error: code = NotFound desc = could not find container \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": container with ID starting with f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.655914 4492 scope.go:117] "RemoveContainer" containerID="171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.656069 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b"} err="failed to get container status \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": rpc error: code = NotFound desc = could not find container \"171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b\": container with ID starting with 171598b56a4733f8ceb72a90ff5fb40b0da845d9e100dc9c79f54f7cfd96b19b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.656100 4492 scope.go:117] "RemoveContainer" containerID="9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.657195 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b"} err="failed to get container status \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": rpc error: code = NotFound desc = could not find container \"9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b\": container with ID starting with 9ad549778d63e9a98894702bffa538bad5fd1f8982ab4ecadd1867b54e01782b not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.657215 4492 scope.go:117] "RemoveContainer" containerID="c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.657595 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704"} err="failed to get container status \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": rpc error: code = NotFound desc = could not find container \"c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704\": container with ID starting with c22c1d8c0d42bcb0e487cdeb8348d5f1434a8897e13062a57b58c8b72d81a704 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.657652 4492 scope.go:117] "RemoveContainer" containerID="f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.658147 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229"} err="failed to get container status \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": rpc error: code = NotFound desc = could not find container \"f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229\": container with ID starting with f2efb421303dd56fd7783e6570b0adf1cd399731349ba80f9cadde28f13d6229 not found: ID does not exist" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.661274 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "3a568bef-11fc-479e-bf90-4c2030405335" (UID: "3a568bef-11fc-479e-bf90-4c2030405335"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.663597 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708094 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-config-data\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708140 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-run-httpd\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708228 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708250 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708335 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-log-httpd\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708380 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn2sh\" (UniqueName: \"kubernetes.io/projected/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-kube-api-access-xn2sh\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708430 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-scripts\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708574 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.708593 4492 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a568bef-11fc-479e-bf90-4c2030405335-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.810776 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-config-data\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.810840 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-run-httpd\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.811942 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-run-httpd\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.811193 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.812016 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.812112 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-log-httpd\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.812492 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn2sh\" (UniqueName: \"kubernetes.io/projected/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-kube-api-access-xn2sh\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.812538 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-scripts\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.812630 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-log-httpd\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.816319 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-config-data\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.817040 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.817187 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-scripts\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.817314 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.837784 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn2sh\" (UniqueName: \"kubernetes.io/projected/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-kube-api-access-xn2sh\") pod \"ceilometer-0\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " pod="openstack/ceilometer-0" Feb 20 06:57:46 crc kubenswrapper[4492]: I0220 06:57:46.976781 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.277732 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c8d654db7-jfjgs" event={"ID":"b14dc392-054d-4257-b340-642e7527aa67","Type":"ContainerStarted","Data":"a2cb1b1641cc3f3d712862760eefb513e26e48d71de226e2ae9c61a1be15cb66"} Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.278063 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c8d654db7-jfjgs" event={"ID":"b14dc392-054d-4257-b340-642e7527aa67","Type":"ContainerStarted","Data":"46a2004aed3e7fe45cbcecab67d2c61890821f788b83f6d9f0119fc22ce7665a"} Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.278141 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c8d654db7-jfjgs" event={"ID":"b14dc392-054d-4257-b340-642e7527aa67","Type":"ContainerStarted","Data":"6acbef1f97d2dc98639d07ef791ea4c78330afb875e7addf55dbd9d5a739fe04"} Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.279120 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.279184 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.287270 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cf7fbb848-5c8g6" event={"ID":"3a568bef-11fc-479e-bf90-4c2030405335","Type":"ContainerDied","Data":"33a80f7ce5836d764d9ad30d5388677e957e6d3468b59372fe4257a99f18aff1"} Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.287335 4492 scope.go:117] "RemoveContainer" containerID="f8b0f7955b6291ba776eb73ad0cea1f5f6d1f8b4ce4d244cf6a302da792fbc0f" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.287538 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cf7fbb848-5c8g6" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.308723 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7c8d654db7-jfjgs" podStartSLOduration=2.308705029 podStartE2EDuration="2.308705029s" podCreationTimestamp="2026-02-20 06:57:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:47.304554517 +0000 UTC m=+1024.075843495" watchObservedRunningTime="2026-02-20 06:57:47.308705029 +0000 UTC m=+1024.079994007" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.339265 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6cf7fbb848-5c8g6"] Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.339889 4492 scope.go:117] "RemoveContainer" containerID="b351db4079da572e45b8ead2fa564d38656849c6e0ce66ea7c1fbcbdb0671376" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.347020 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6cf7fbb848-5c8g6"] Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.446631 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.570849 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0318c0cb-0268-4818-872b-b9a194dc0622" path="/var/lib/kubelet/pods/0318c0cb-0268-4818-872b-b9a194dc0622/volumes" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.571594 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a568bef-11fc-479e-bf90-4c2030405335" path="/var/lib/kubelet/pods/3a568bef-11fc-479e-bf90-4c2030405335/volumes" Feb 20 06:57:47 crc kubenswrapper[4492]: I0220 06:57:47.874501 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 20 06:57:48 crc kubenswrapper[4492]: I0220 06:57:48.302967 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerStarted","Data":"4d886fb44ac08fd6e37b85ef29a5a988821ce0bd19239868fa5787f43e81268d"} Feb 20 06:57:49 crc kubenswrapper[4492]: I0220 06:57:49.313592 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerStarted","Data":"23b25f0ddbab7f6803ecc13c5e0c9b0aebf22d7f3c1bf84aac2dcc3ba90297e1"} Feb 20 06:57:49 crc kubenswrapper[4492]: I0220 06:57:49.313870 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerStarted","Data":"c66c7d455e6be216939439401d091ce8dd637fd073bd71c7aa443efca31619be"} Feb 20 06:57:50 crc kubenswrapper[4492]: I0220 06:57:50.348026 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerStarted","Data":"288368ed8b45752696739a4b85c3bcd829998eec79d3edd3d7ed060f74d30e08"} Feb 20 06:57:50 crc kubenswrapper[4492]: I0220 06:57:50.932719 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-58b9894566-r5jqg"] Feb 20 06:57:50 crc kubenswrapper[4492]: I0220 06:57:50.934345 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:50 crc kubenswrapper[4492]: I0220 06:57:50.940825 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Feb 20 06:57:50 crc kubenswrapper[4492]: I0220 06:57:50.941075 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-5wknk" Feb 20 06:57:50 crc kubenswrapper[4492]: I0220 06:57:50.941328 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Feb 20 06:57:50 crc kubenswrapper[4492]: I0220 06:57:50.963051 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-58b9894566-r5jqg"] Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.035743 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data-custom\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.037576 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.037658 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-combined-ca-bundle\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.037715 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7grtm\" (UniqueName: \"kubernetes.io/projected/e7da01e6-bb65-47dd-8357-72f1fec9d95d-kube-api-access-7grtm\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.139928 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-55d4487b7d-v89bg"] Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.153703 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-combined-ca-bundle\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.153883 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7grtm\" (UniqueName: \"kubernetes.io/projected/e7da01e6-bb65-47dd-8357-72f1fec9d95d-kube-api-access-7grtm\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.154081 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data-custom\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.154117 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.158067 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.161225 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.187997 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-55d4487b7d-v89bg"] Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.204298 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data-custom\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.257160 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.257248 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vmwc\" (UniqueName: \"kubernetes.io/projected/cd038612-82db-4064-9515-9f4563121d37-kube-api-access-9vmwc\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.257360 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data-custom\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.257551 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-combined-ca-bundle\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.259834 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-combined-ca-bundle\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.260713 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7grtm\" (UniqueName: \"kubernetes.io/projected/e7da01e6-bb65-47dd-8357-72f1fec9d95d-kube-api-access-7grtm\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.270110 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7558f88fd7-hk2tb"] Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.271195 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65f5db898d-bxntp" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.286818 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.288165 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data\") pod \"heat-engine-58b9894566-r5jqg\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.298103 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7558f88fd7-hk2tb"] Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.334091 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.341795 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6cd457fdb6-qlxrk"] Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.343294 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.346635 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.367172 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6cd457fdb6-qlxrk"] Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.368603 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-nb\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.368633 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data-custom\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.368704 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4w7h\" (UniqueName: \"kubernetes.io/projected/e57e76c2-a16c-43a6-96c0-6b891618e5c0-kube-api-access-w4w7h\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.368734 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-swift-storage-0\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.368975 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-sb\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369003 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data-custom\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369021 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-combined-ca-bundle\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369153 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-combined-ca-bundle\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369194 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hjns\" (UniqueName: \"kubernetes.io/projected/aae0abff-d540-481e-9b6d-b3e745c89db2-kube-api-access-2hjns\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369231 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369279 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vmwc\" (UniqueName: \"kubernetes.io/projected/cd038612-82db-4064-9515-9f4563121d37-kube-api-access-9vmwc\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369443 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369507 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-svc\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.369530 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-config\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.376446 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data-custom\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.376544 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-combined-ca-bundle\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.386696 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.389334 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vmwc\" (UniqueName: \"kubernetes.io/projected/cd038612-82db-4064-9515-9f4563121d37-kube-api-access-9vmwc\") pod \"heat-cfnapi-55d4487b7d-v89bg\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.389514 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerStarted","Data":"6b658b53be664062d0337d6b6fc83749327873afca9013f7a0a14b9450478cae"} Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.390692 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.422257 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.994312675 podStartE2EDuration="5.422237439s" podCreationTimestamp="2026-02-20 06:57:46 +0000 UTC" firstStartedPulling="2026-02-20 06:57:47.449209017 +0000 UTC m=+1024.220497995" lastFinishedPulling="2026-02-20 06:57:50.877133781 +0000 UTC m=+1027.648422759" observedRunningTime="2026-02-20 06:57:51.41697949 +0000 UTC m=+1028.188268468" watchObservedRunningTime="2026-02-20 06:57:51.422237439 +0000 UTC m=+1028.193526418" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471267 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hjns\" (UniqueName: \"kubernetes.io/projected/aae0abff-d540-481e-9b6d-b3e745c89db2-kube-api-access-2hjns\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471367 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471403 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-svc\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471422 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-config\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471441 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-nb\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471496 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4w7h\" (UniqueName: \"kubernetes.io/projected/e57e76c2-a16c-43a6-96c0-6b891618e5c0-kube-api-access-w4w7h\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471519 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-swift-storage-0\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471597 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-sb\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471615 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data-custom\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.471632 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-combined-ca-bundle\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.476156 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-swift-storage-0\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.477502 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-nb\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.477509 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-sb\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.479095 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-svc\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.479256 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-config\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.484298 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.491243 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-combined-ca-bundle\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.491955 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data-custom\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.496294 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hjns\" (UniqueName: \"kubernetes.io/projected/aae0abff-d540-481e-9b6d-b3e745c89db2-kube-api-access-2hjns\") pod \"dnsmasq-dns-7558f88fd7-hk2tb\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.499633 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4w7h\" (UniqueName: \"kubernetes.io/projected/e57e76c2-a16c-43a6-96c0-6b891618e5c0-kube-api-access-w4w7h\") pod \"heat-api-6cd457fdb6-qlxrk\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.612644 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.635099 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.702804 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:57:51 crc kubenswrapper[4492]: I0220 06:57:51.847260 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-58b9894566-r5jqg"] Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.166280 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7558f88fd7-hk2tb"] Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.277820 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-55d4487b7d-v89bg"] Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.338842 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6cd457fdb6-qlxrk"] Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.409646 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" event={"ID":"aae0abff-d540-481e-9b6d-b3e745c89db2","Type":"ContainerStarted","Data":"a44c83098f25b719e10d5f8cd22b596be017689372747c75ce0d2b9435fc183d"} Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.416724 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" event={"ID":"cd038612-82db-4064-9515-9f4563121d37","Type":"ContainerStarted","Data":"ccfba3bac889921de8993172e3c8abdf1aaac7094cbc07e106914323ecd28945"} Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.423560 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6cd457fdb6-qlxrk" event={"ID":"e57e76c2-a16c-43a6-96c0-6b891618e5c0","Type":"ContainerStarted","Data":"2a65f27167d6eb2c38e2c78cbd55261156bb30c414aad9b7bd9f1d9c2474f4b0"} Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.431573 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-58b9894566-r5jqg" event={"ID":"e7da01e6-bb65-47dd-8357-72f1fec9d95d","Type":"ContainerStarted","Data":"8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e"} Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.431607 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-58b9894566-r5jqg" event={"ID":"e7da01e6-bb65-47dd-8357-72f1fec9d95d","Type":"ContainerStarted","Data":"68432b00028a0d500ea5e2fb5eb7e74a949405f5c23055ddd51370e876bc2db1"} Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.431644 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:57:52 crc kubenswrapper[4492]: I0220 06:57:52.452920 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-58b9894566-r5jqg" podStartSLOduration=2.45290304 podStartE2EDuration="2.45290304s" podCreationTimestamp="2026-02-20 06:57:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:52.448906116 +0000 UTC m=+1029.220195095" watchObservedRunningTime="2026-02-20 06:57:52.45290304 +0000 UTC m=+1029.224192018" Feb 20 06:57:53 crc kubenswrapper[4492]: I0220 06:57:53.468342 4492 generic.go:334] "Generic (PLEG): container finished" podID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerID="ae1beeca154e35d9335bc9f4eda8f838567b3fc60fbd3ae3445bbc7b6ef851bc" exitCode=0 Feb 20 06:57:53 crc kubenswrapper[4492]: I0220 06:57:53.468766 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" event={"ID":"aae0abff-d540-481e-9b6d-b3e745c89db2","Type":"ContainerDied","Data":"ae1beeca154e35d9335bc9f4eda8f838567b3fc60fbd3ae3445bbc7b6ef851bc"} Feb 20 06:57:54 crc kubenswrapper[4492]: E0220 06:57:54.429500 4492 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/9545fbd143472fc8b8f9ec0b8d0b5e6f080eae119701b2aa1938d343d60d8501/diff" to get inode usage: stat /var/lib/containers/storage/overlay/9545fbd143472fc8b8f9ec0b8d0b5e6f080eae119701b2aa1938d343d60d8501/diff: no such file or directory, extraDiskErr: Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.484785 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" event={"ID":"aae0abff-d540-481e-9b6d-b3e745c89db2","Type":"ContainerStarted","Data":"e704c453c7f3c513c689ecd8855fb2b74069fbf0a5f19a792326dd662dbde1d4"} Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.484837 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.504722 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" podStartSLOduration=3.504701102 podStartE2EDuration="3.504701102s" podCreationTimestamp="2026-02-20 06:57:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:57:54.497902638 +0000 UTC m=+1031.269191616" watchObservedRunningTime="2026-02-20 06:57:54.504701102 +0000 UTC m=+1031.275990079" Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.884068 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.884535 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-central-agent" containerID="cri-o://c66c7d455e6be216939439401d091ce8dd637fd073bd71c7aa443efca31619be" gracePeriod=30 Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.884560 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="proxy-httpd" containerID="cri-o://6b658b53be664062d0337d6b6fc83749327873afca9013f7a0a14b9450478cae" gracePeriod=30 Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.884611 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="sg-core" containerID="cri-o://288368ed8b45752696739a4b85c3bcd829998eec79d3edd3d7ed060f74d30e08" gracePeriod=30 Feb 20 06:57:54 crc kubenswrapper[4492]: I0220 06:57:54.884626 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-notification-agent" containerID="cri-o://23b25f0ddbab7f6803ecc13c5e0c9b0aebf22d7f3c1bf84aac2dcc3ba90297e1" gracePeriod=30 Feb 20 06:57:55 crc kubenswrapper[4492]: E0220 06:57:55.221047 4492 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/5392b97afe6a183001424bfd49acc272b447bbdbf441955b4b6838a3322c729f/diff" to get inode usage: stat /var/lib/containers/storage/overlay/5392b97afe6a183001424bfd49acc272b447bbdbf441955b4b6838a3322c729f/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-6cf7fbb848-5c8g6_3a568bef-11fc-479e-bf90-4c2030405335/neutron-api/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-6cf7fbb848-5c8g6_3a568bef-11fc-479e-bf90-4c2030405335/neutron-api/0.log: no such file or directory Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.497816 4492 generic.go:334] "Generic (PLEG): container finished" podID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerID="6b658b53be664062d0337d6b6fc83749327873afca9013f7a0a14b9450478cae" exitCode=0 Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.497863 4492 generic.go:334] "Generic (PLEG): container finished" podID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerID="288368ed8b45752696739a4b85c3bcd829998eec79d3edd3d7ed060f74d30e08" exitCode=2 Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.497872 4492 generic.go:334] "Generic (PLEG): container finished" podID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerID="23b25f0ddbab7f6803ecc13c5e0c9b0aebf22d7f3c1bf84aac2dcc3ba90297e1" exitCode=0 Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.498637 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerDied","Data":"6b658b53be664062d0337d6b6fc83749327873afca9013f7a0a14b9450478cae"} Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.498686 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerDied","Data":"288368ed8b45752696739a4b85c3bcd829998eec79d3edd3d7ed060f74d30e08"} Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.498698 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerDied","Data":"23b25f0ddbab7f6803ecc13c5e0c9b0aebf22d7f3c1bf84aac2dcc3ba90297e1"} Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.711154 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:55 crc kubenswrapper[4492]: I0220 06:57:55.712011 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c8d654db7-jfjgs" Feb 20 06:57:56 crc kubenswrapper[4492]: I0220 06:57:56.527579 4492 generic.go:334] "Generic (PLEG): container finished" podID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerID="c66c7d455e6be216939439401d091ce8dd637fd073bd71c7aa443efca31619be" exitCode=0 Feb 20 06:57:56 crc kubenswrapper[4492]: I0220 06:57:56.527683 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerDied","Data":"c66c7d455e6be216939439401d091ce8dd637fd073bd71c7aa443efca31619be"} Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.826465 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-749c7585d6-zr8k4"] Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.832412 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.883558 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-749c7585d6-zr8k4"] Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.894303 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5d6f59fdc-lr7g6"] Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.896569 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.921228 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-8c46948df-55k9j"] Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.922758 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.935164 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-config-data\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.935219 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzj64\" (UniqueName: \"kubernetes.io/projected/618181e6-1990-44c6-b7c5-389d5609252d-kube-api-access-wzj64\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.935354 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-config-data-custom\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.935376 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-combined-ca-bundle\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.943465 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5d6f59fdc-lr7g6"] Feb 20 06:57:58 crc kubenswrapper[4492]: I0220 06:57:58.951422 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8c46948df-55k9j"] Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.037983 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038041 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038070 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-combined-ca-bundle\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038131 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data-custom\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038156 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgt4q\" (UniqueName: \"kubernetes.io/projected/a789a142-4568-4e21-a16e-75749835da10-kube-api-access-xgt4q\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038185 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-config-data\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038214 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzj64\" (UniqueName: \"kubernetes.io/projected/618181e6-1990-44c6-b7c5-389d5609252d-kube-api-access-wzj64\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038249 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-combined-ca-bundle\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038270 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data-custom\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038299 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9tmc\" (UniqueName: \"kubernetes.io/projected/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-kube-api-access-n9tmc\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038323 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-config-data-custom\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.038343 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-combined-ca-bundle\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.049227 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-config-data\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.051736 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-combined-ca-bundle\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.059123 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/618181e6-1990-44c6-b7c5-389d5609252d-config-data-custom\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.066506 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzj64\" (UniqueName: \"kubernetes.io/projected/618181e6-1990-44c6-b7c5-389d5609252d-kube-api-access-wzj64\") pod \"heat-engine-749c7585d6-zr8k4\" (UID: \"618181e6-1990-44c6-b7c5-389d5609252d\") " pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.141489 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data-custom\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.141566 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgt4q\" (UniqueName: \"kubernetes.io/projected/a789a142-4568-4e21-a16e-75749835da10-kube-api-access-xgt4q\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.141658 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-combined-ca-bundle\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.141707 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data-custom\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.141763 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9tmc\" (UniqueName: \"kubernetes.io/projected/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-kube-api-access-n9tmc\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.144996 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.145647 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.145723 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-combined-ca-bundle\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.148185 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data-custom\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.156375 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.156826 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data-custom\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.159906 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.161301 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgt4q\" (UniqueName: \"kubernetes.io/projected/a789a142-4568-4e21-a16e-75749835da10-kube-api-access-xgt4q\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.161404 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-combined-ca-bundle\") pod \"heat-api-5d6f59fdc-lr7g6\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.162386 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-combined-ca-bundle\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.162656 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9tmc\" (UniqueName: \"kubernetes.io/projected/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-kube-api-access-n9tmc\") pod \"heat-cfnapi-8c46948df-55k9j\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.166802 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.217997 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:57:59 crc kubenswrapper[4492]: I0220 06:57:59.251179 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.110898 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6cd457fdb6-qlxrk"] Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.126855 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-55d4487b7d-v89bg"] Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.132971 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-d44b9d589-wkjn4"] Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.135363 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.144821 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.145048 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.146491 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-d44b9d589-wkjn4"] Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.177826 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-86b9576f68-82n82"] Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.180151 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.196110 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.196448 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.202865 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86b9576f68-82n82"] Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.295354 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-combined-ca-bundle\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.295507 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-config-data\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.295556 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-public-tls-certs\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296368 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-internal-tls-certs\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296425 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-public-tls-certs\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296456 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-config-data-custom\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296555 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-config-data-custom\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296618 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6c65\" (UniqueName: \"kubernetes.io/projected/188920e4-7af8-4821-9428-d6224cfeba8e-kube-api-access-z6c65\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296671 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-config-data\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296750 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-combined-ca-bundle\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296815 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-internal-tls-certs\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.296944 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcjm9\" (UniqueName: \"kubernetes.io/projected/270094d8-faff-42ce-9a33-3f9687129320-kube-api-access-vcjm9\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.399950 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-config-data-custom\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.399999 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-config-data-custom\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400027 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6c65\" (UniqueName: \"kubernetes.io/projected/188920e4-7af8-4821-9428-d6224cfeba8e-kube-api-access-z6c65\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400051 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-config-data\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400088 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-combined-ca-bundle\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400113 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-internal-tls-certs\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400165 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcjm9\" (UniqueName: \"kubernetes.io/projected/270094d8-faff-42ce-9a33-3f9687129320-kube-api-access-vcjm9\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400188 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-combined-ca-bundle\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400230 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-config-data\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400262 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-public-tls-certs\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400290 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-internal-tls-certs\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.400311 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-public-tls-certs\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.407597 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-combined-ca-bundle\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.408584 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-public-tls-certs\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.408689 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-config-data-custom\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.412252 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-internal-tls-certs\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.415516 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-combined-ca-bundle\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.416992 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-config-data-custom\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.419072 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-public-tls-certs\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.424308 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-internal-tls-certs\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.432539 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188920e4-7af8-4821-9428-d6224cfeba8e-config-data\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.434851 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcjm9\" (UniqueName: \"kubernetes.io/projected/270094d8-faff-42ce-9a33-3f9687129320-kube-api-access-vcjm9\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.435119 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6c65\" (UniqueName: \"kubernetes.io/projected/188920e4-7af8-4821-9428-d6224cfeba8e-kube-api-access-z6c65\") pod \"heat-api-d44b9d589-wkjn4\" (UID: \"188920e4-7af8-4821-9428-d6224cfeba8e\") " pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.435376 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270094d8-faff-42ce-9a33-3f9687129320-config-data\") pod \"heat-cfnapi-86b9576f68-82n82\" (UID: \"270094d8-faff-42ce-9a33-3f9687129320\") " pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.466299 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:00 crc kubenswrapper[4492]: I0220 06:58:00.517482 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:01 crc kubenswrapper[4492]: I0220 06:58:01.267244 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-65f5db898d-bxntp" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Feb 20 06:58:01 crc kubenswrapper[4492]: I0220 06:58:01.267764 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:58:01 crc kubenswrapper[4492]: I0220 06:58:01.636685 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:58:01 crc kubenswrapper[4492]: I0220 06:58:01.715328 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d4fbd665-cbckt"] Feb 20 06:58:01 crc kubenswrapper[4492]: I0220 06:58:01.725357 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" podUID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerName="dnsmasq-dns" containerID="cri-o://2ce5643fc51d0db7366057f9d56f5d38c433a025228ec00329e5e4047cc3d2f7" gracePeriod=10 Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.603555 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b","Type":"ContainerDied","Data":"4d886fb44ac08fd6e37b85ef29a5a988821ce0bd19239868fa5787f43e81268d"} Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.603882 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d886fb44ac08fd6e37b85ef29a5a988821ce0bd19239868fa5787f43e81268d" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.627205 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.629833 4492 generic.go:334] "Generic (PLEG): container finished" podID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerID="2ce5643fc51d0db7366057f9d56f5d38c433a025228ec00329e5e4047cc3d2f7" exitCode=0 Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.629878 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" event={"ID":"8bdbc0d0-444f-495f-bd72-57c41c6bcd71","Type":"ContainerDied","Data":"2ce5643fc51d0db7366057f9d56f5d38c433a025228ec00329e5e4047cc3d2f7"} Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.772985 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-config-data\") pod \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.773340 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-combined-ca-bundle\") pod \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.773463 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-log-httpd\") pod \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.773599 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-sg-core-conf-yaml\") pod \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.773700 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-run-httpd\") pod \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.773820 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-scripts\") pod \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.773866 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn2sh\" (UniqueName: \"kubernetes.io/projected/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-kube-api-access-xn2sh\") pod \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\" (UID: \"7b766f8e-6746-48b9-8bc9-3d877bb8cf1b\") " Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.773961 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" (UID: "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.774759 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" (UID: "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.777997 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.778022 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.782883 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-scripts" (OuterVolumeSpecName: "scripts") pod "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" (UID: "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.791694 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-kube-api-access-xn2sh" (OuterVolumeSpecName: "kube-api-access-xn2sh") pod "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" (UID: "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b"). InnerVolumeSpecName "kube-api-access-xn2sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.876109 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" (UID: "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.900360 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn2sh\" (UniqueName: \"kubernetes.io/projected/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-kube-api-access-xn2sh\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.900380 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.900390 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:02 crc kubenswrapper[4492]: I0220 06:58:02.949556 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" (UID: "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.002108 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.029117 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-config-data" (OuterVolumeSpecName: "config-data") pod "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" (UID: "7b766f8e-6746-48b9-8bc9-3d877bb8cf1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.104308 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.108580 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.205865 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tcth\" (UniqueName: \"kubernetes.io/projected/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-kube-api-access-5tcth\") pod \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.205949 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-sb\") pod \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.206056 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-svc\") pod \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.206085 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-swift-storage-0\") pod \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.206325 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-nb\") pod \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.206421 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-config\") pod \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\" (UID: \"8bdbc0d0-444f-495f-bd72-57c41c6bcd71\") " Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.232637 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8c46948df-55k9j"] Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.238451 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-kube-api-access-5tcth" (OuterVolumeSpecName: "kube-api-access-5tcth") pod "8bdbc0d0-444f-495f-bd72-57c41c6bcd71" (UID: "8bdbc0d0-444f-495f-bd72-57c41c6bcd71"). InnerVolumeSpecName "kube-api-access-5tcth". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.268941 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86b9576f68-82n82"] Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.309320 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tcth\" (UniqueName: \"kubernetes.io/projected/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-kube-api-access-5tcth\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.353058 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-config" (OuterVolumeSpecName: "config") pod "8bdbc0d0-444f-495f-bd72-57c41c6bcd71" (UID: "8bdbc0d0-444f-495f-bd72-57c41c6bcd71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.372746 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-749c7585d6-zr8k4"] Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.376055 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8bdbc0d0-444f-495f-bd72-57c41c6bcd71" (UID: "8bdbc0d0-444f-495f-bd72-57c41c6bcd71"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.376571 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8bdbc0d0-444f-495f-bd72-57c41c6bcd71" (UID: "8bdbc0d0-444f-495f-bd72-57c41c6bcd71"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.383192 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8bdbc0d0-444f-495f-bd72-57c41c6bcd71" (UID: "8bdbc0d0-444f-495f-bd72-57c41c6bcd71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.388185 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5d6f59fdc-lr7g6"] Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.412315 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.412893 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.412940 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.412951 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.414993 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8bdbc0d0-444f-495f-bd72-57c41c6bcd71" (UID: "8bdbc0d0-444f-495f-bd72-57c41c6bcd71"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.434824 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-d44b9d589-wkjn4"] Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.518302 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8bdbc0d0-444f-495f-bd72-57c41c6bcd71-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.658139 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-749c7585d6-zr8k4" event={"ID":"618181e6-1990-44c6-b7c5-389d5609252d","Type":"ContainerStarted","Data":"b6dd03a9d9ab5fb0af91a20f073fb9fcc5365ca79992458fb06947d5f12dbaf7"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.658186 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-749c7585d6-zr8k4" event={"ID":"618181e6-1990-44c6-b7c5-389d5609252d","Type":"ContainerStarted","Data":"69bd232cd68eb9927d4a0138358b8f7df35f3383f423a0c5a541d4729a54f670"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.659265 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.667365 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" event={"ID":"8bdbc0d0-444f-495f-bd72-57c41c6bcd71","Type":"ContainerDied","Data":"6592d07ab3200040ce1380a5185adb60d9d213deaf37653b2f5645919867d4a9"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.667453 4492 scope.go:117] "RemoveContainer" containerID="2ce5643fc51d0db7366057f9d56f5d38c433a025228ec00329e5e4047cc3d2f7" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.667602 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d4fbd665-cbckt" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.684053 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c46948df-55k9j" event={"ID":"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392","Type":"ContainerStarted","Data":"4e1a85ff90fbd67d56f4fb7e4f1f68b12f035e4799df9cca07f88a3425f7a941"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.684084 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c46948df-55k9j" event={"ID":"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392","Type":"ContainerStarted","Data":"03214e31f63648e413d7f5d334a5b4143fcc782b13710dea8f2db5fe13863968"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.684552 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.699594 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d6f59fdc-lr7g6" event={"ID":"a789a142-4568-4e21-a16e-75749835da10","Type":"ContainerStarted","Data":"598d455259a6b8b75540134eb35c7efe5d3f20c671aa5f5c449f3f828a41ec51"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.711741 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d44b9d589-wkjn4" event={"ID":"188920e4-7af8-4821-9428-d6224cfeba8e","Type":"ContainerStarted","Data":"2168df7223b366d830e834afcf5bfac3f1efd26d0513c5b67fd8df04c6f0a990"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.749812 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b9576f68-82n82" event={"ID":"270094d8-faff-42ce-9a33-3f9687129320","Type":"ContainerStarted","Data":"473b9e3c26fb7e7cdb7809a422c2548c5544a384ad02d1c1c08a1901b4833505"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.749872 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b9576f68-82n82" event={"ID":"270094d8-faff-42ce-9a33-3f9687129320","Type":"ContainerStarted","Data":"fec54065c8df4b5010d4963b6f6e73008d2d4762571da18ed21a98de4173f9a9"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.750893 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.770066 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" event={"ID":"cd038612-82db-4064-9515-9f4563121d37","Type":"ContainerStarted","Data":"d6a493a97cb0c02bb235a48fc228b9bc57e346d36c7210589dd10e7d2c3f7c20"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.770213 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" podUID="cd038612-82db-4064-9515-9f4563121d37" containerName="heat-cfnapi" containerID="cri-o://d6a493a97cb0c02bb235a48fc228b9bc57e346d36c7210589dd10e7d2c3f7c20" gracePeriod=60 Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.770429 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.795611 4492 scope.go:117] "RemoveContainer" containerID="06beb6d9fd8852edc6ee30c9dde953a2b48c8c56c5915464977e08077d7ef9d1" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.803776 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6cd457fdb6-qlxrk" event={"ID":"e57e76c2-a16c-43a6-96c0-6b891618e5c0","Type":"ContainerStarted","Data":"20fbefa0675a7b6dcc5cb14869b50309451cb5a576d8e7d85a582d7349fc6482"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.803916 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6cd457fdb6-qlxrk" podUID="e57e76c2-a16c-43a6-96c0-6b891618e5c0" containerName="heat-api" containerID="cri-o://20fbefa0675a7b6dcc5cb14869b50309451cb5a576d8e7d85a582d7349fc6482" gracePeriod=60 Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.804151 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.823523 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d4fbd665-cbckt"] Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.841813 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.842193 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"aa29dc13-c6f5-43ef-b0a9-44b6735922d7","Type":"ContainerStarted","Data":"3f548c3844d86045d10c8047e7d7c91f48020b82f9185eec0e98e7d499f23aa4"} Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.843838 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d4fbd665-cbckt"] Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.844924 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-749c7585d6-zr8k4" podStartSLOduration=5.844913736 podStartE2EDuration="5.844913736s" podCreationTimestamp="2026-02-20 06:57:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:03.804748774 +0000 UTC m=+1040.576037753" watchObservedRunningTime="2026-02-20 06:58:03.844913736 +0000 UTC m=+1040.616202714" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.858915 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-8c46948df-55k9j" podStartSLOduration=5.858906472 podStartE2EDuration="5.858906472s" podCreationTimestamp="2026-02-20 06:57:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:03.829611341 +0000 UTC m=+1040.600900319" watchObservedRunningTime="2026-02-20 06:58:03.858906472 +0000 UTC m=+1040.630195451" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.879774 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" podStartSLOduration=2.6177719760000002 podStartE2EDuration="12.879757287s" podCreationTimestamp="2026-02-20 06:57:51 +0000 UTC" firstStartedPulling="2026-02-20 06:57:52.273088161 +0000 UTC m=+1029.044377139" lastFinishedPulling="2026-02-20 06:58:02.535073471 +0000 UTC m=+1039.306362450" observedRunningTime="2026-02-20 06:58:03.863417055 +0000 UTC m=+1040.634706033" watchObservedRunningTime="2026-02-20 06:58:03.879757287 +0000 UTC m=+1040.651046265" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.898941 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-86b9576f68-82n82" podStartSLOduration=3.898930438 podStartE2EDuration="3.898930438s" podCreationTimestamp="2026-02-20 06:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:03.895038913 +0000 UTC m=+1040.666327891" watchObservedRunningTime="2026-02-20 06:58:03.898930438 +0000 UTC m=+1040.670219416" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.934700 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6cd457fdb6-qlxrk" podStartSLOduration=2.799656989 podStartE2EDuration="12.934685057s" podCreationTimestamp="2026-02-20 06:57:51 +0000 UTC" firstStartedPulling="2026-02-20 06:57:52.384253054 +0000 UTC m=+1029.155542033" lastFinishedPulling="2026-02-20 06:58:02.519281133 +0000 UTC m=+1039.290570101" observedRunningTime="2026-02-20 06:58:03.916465243 +0000 UTC m=+1040.687754221" watchObservedRunningTime="2026-02-20 06:58:03.934685057 +0000 UTC m=+1040.705974035" Feb 20 06:58:03 crc kubenswrapper[4492]: I0220 06:58:03.987170 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.043395 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.057794 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.521610977 podStartE2EDuration="20.057759807s" podCreationTimestamp="2026-02-20 06:57:44 +0000 UTC" firstStartedPulling="2026-02-20 06:57:44.953455611 +0000 UTC m=+1021.724744590" lastFinishedPulling="2026-02-20 06:58:02.489604441 +0000 UTC m=+1039.260893420" observedRunningTime="2026-02-20 06:58:03.954351399 +0000 UTC m=+1040.725640377" watchObservedRunningTime="2026-02-20 06:58:04.057759807 +0000 UTC m=+1040.829048784" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.083535 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:04 crc kubenswrapper[4492]: E0220 06:58:04.083981 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="proxy-httpd" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084003 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="proxy-httpd" Feb 20 06:58:04 crc kubenswrapper[4492]: E0220 06:58:04.084019 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-notification-agent" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084027 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-notification-agent" Feb 20 06:58:04 crc kubenswrapper[4492]: E0220 06:58:04.084039 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="sg-core" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084045 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="sg-core" Feb 20 06:58:04 crc kubenswrapper[4492]: E0220 06:58:04.084055 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerName="init" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084061 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerName="init" Feb 20 06:58:04 crc kubenswrapper[4492]: E0220 06:58:04.084077 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-central-agent" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084083 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-central-agent" Feb 20 06:58:04 crc kubenswrapper[4492]: E0220 06:58:04.084094 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerName="dnsmasq-dns" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084100 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerName="dnsmasq-dns" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084298 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" containerName="dnsmasq-dns" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084316 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-notification-agent" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084332 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="proxy-httpd" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084341 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="ceilometer-central-agent" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.084349 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" containerName="sg-core" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.086150 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.086245 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.089965 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.090709 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.240580 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-config-data\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.240618 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-run-httpd\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.240756 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-scripts\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.240821 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcppk\" (UniqueName: \"kubernetes.io/projected/3f9c0e03-bed9-41e5-86c2-633bcf86632f-kube-api-access-zcppk\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.240893 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.240929 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-log-httpd\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.240989 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.342623 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-log-httpd\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.342700 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.342739 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-config-data\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.342754 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-run-httpd\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.342813 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-scripts\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.342849 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcppk\" (UniqueName: \"kubernetes.io/projected/3f9c0e03-bed9-41e5-86c2-633bcf86632f-kube-api-access-zcppk\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.342887 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.348561 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-run-httpd\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.348605 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-log-httpd\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.351285 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-config-data\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.352870 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.357933 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-scripts\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.359002 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.369319 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcppk\" (UniqueName: \"kubernetes.io/projected/3f9c0e03-bed9-41e5-86c2-633bcf86632f-kube-api-access-zcppk\") pod \"ceilometer-0\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.408231 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.858895 4492 generic.go:334] "Generic (PLEG): container finished" podID="e57e76c2-a16c-43a6-96c0-6b891618e5c0" containerID="20fbefa0675a7b6dcc5cb14869b50309451cb5a576d8e7d85a582d7349fc6482" exitCode=0 Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.858957 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6cd457fdb6-qlxrk" event={"ID":"e57e76c2-a16c-43a6-96c0-6b891618e5c0","Type":"ContainerDied","Data":"20fbefa0675a7b6dcc5cb14869b50309451cb5a576d8e7d85a582d7349fc6482"} Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.862930 4492 generic.go:334] "Generic (PLEG): container finished" podID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerID="4e1a85ff90fbd67d56f4fb7e4f1f68b12f035e4799df9cca07f88a3425f7a941" exitCode=1 Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.863029 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c46948df-55k9j" event={"ID":"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392","Type":"ContainerDied","Data":"4e1a85ff90fbd67d56f4fb7e4f1f68b12f035e4799df9cca07f88a3425f7a941"} Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.866346 4492 generic.go:334] "Generic (PLEG): container finished" podID="a789a142-4568-4e21-a16e-75749835da10" containerID="27fb0a610089d7a59cca6ac577c161ee12efa632b8c1120837bb874837d1d5ea" exitCode=1 Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.866699 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d6f59fdc-lr7g6" event={"ID":"a789a142-4568-4e21-a16e-75749835da10","Type":"ContainerDied","Data":"27fb0a610089d7a59cca6ac577c161ee12efa632b8c1120837bb874837d1d5ea"} Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.867865 4492 scope.go:117] "RemoveContainer" containerID="4e1a85ff90fbd67d56f4fb7e4f1f68b12f035e4799df9cca07f88a3425f7a941" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.868791 4492 scope.go:117] "RemoveContainer" containerID="27fb0a610089d7a59cca6ac577c161ee12efa632b8c1120837bb874837d1d5ea" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.877399 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d44b9d589-wkjn4" event={"ID":"188920e4-7af8-4821-9428-d6224cfeba8e","Type":"ContainerStarted","Data":"bdf7927674a7c0268b2ba36839fdcde01ddba480d14cc4b0b46cc5662789f334"} Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.878078 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.889632 4492 generic.go:334] "Generic (PLEG): container finished" podID="cd038612-82db-4064-9515-9f4563121d37" containerID="d6a493a97cb0c02bb235a48fc228b9bc57e346d36c7210589dd10e7d2c3f7c20" exitCode=0 Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.889704 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" event={"ID":"cd038612-82db-4064-9515-9f4563121d37","Type":"ContainerDied","Data":"d6a493a97cb0c02bb235a48fc228b9bc57e346d36c7210589dd10e7d2c3f7c20"} Feb 20 06:58:04 crc kubenswrapper[4492]: I0220 06:58:04.976858 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-d44b9d589-wkjn4" podStartSLOduration=4.976825761 podStartE2EDuration="4.976825761s" podCreationTimestamp="2026-02-20 06:58:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:04.933647799 +0000 UTC m=+1041.704936777" watchObservedRunningTime="2026-02-20 06:58:04.976825761 +0000 UTC m=+1041.748114739" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.011384 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.140151 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.146025 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.267244 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data-custom\") pod \"cd038612-82db-4064-9515-9f4563121d37\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.267714 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data-custom\") pod \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.267793 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vmwc\" (UniqueName: \"kubernetes.io/projected/cd038612-82db-4064-9515-9f4563121d37-kube-api-access-9vmwc\") pod \"cd038612-82db-4064-9515-9f4563121d37\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.267822 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-combined-ca-bundle\") pod \"cd038612-82db-4064-9515-9f4563121d37\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.267978 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-combined-ca-bundle\") pod \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.268130 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data\") pod \"cd038612-82db-4064-9515-9f4563121d37\" (UID: \"cd038612-82db-4064-9515-9f4563121d37\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.268263 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data\") pod \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.268302 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4w7h\" (UniqueName: \"kubernetes.io/projected/e57e76c2-a16c-43a6-96c0-6b891618e5c0-kube-api-access-w4w7h\") pod \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\" (UID: \"e57e76c2-a16c-43a6-96c0-6b891618e5c0\") " Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.275664 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd038612-82db-4064-9515-9f4563121d37-kube-api-access-9vmwc" (OuterVolumeSpecName: "kube-api-access-9vmwc") pod "cd038612-82db-4064-9515-9f4563121d37" (UID: "cd038612-82db-4064-9515-9f4563121d37"). InnerVolumeSpecName "kube-api-access-9vmwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.275860 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e57e76c2-a16c-43a6-96c0-6b891618e5c0" (UID: "e57e76c2-a16c-43a6-96c0-6b891618e5c0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.276815 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cd038612-82db-4064-9515-9f4563121d37" (UID: "cd038612-82db-4064-9515-9f4563121d37"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.278658 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e57e76c2-a16c-43a6-96c0-6b891618e5c0-kube-api-access-w4w7h" (OuterVolumeSpecName: "kube-api-access-w4w7h") pod "e57e76c2-a16c-43a6-96c0-6b891618e5c0" (UID: "e57e76c2-a16c-43a6-96c0-6b891618e5c0"). InnerVolumeSpecName "kube-api-access-w4w7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.317637 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e57e76c2-a16c-43a6-96c0-6b891618e5c0" (UID: "e57e76c2-a16c-43a6-96c0-6b891618e5c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.320187 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd038612-82db-4064-9515-9f4563121d37" (UID: "cd038612-82db-4064-9515-9f4563121d37"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.349203 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data" (OuterVolumeSpecName: "config-data") pod "e57e76c2-a16c-43a6-96c0-6b891618e5c0" (UID: "e57e76c2-a16c-43a6-96c0-6b891618e5c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.353542 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data" (OuterVolumeSpecName: "config-data") pod "cd038612-82db-4064-9515-9f4563121d37" (UID: "cd038612-82db-4064-9515-9f4563121d37"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370573 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370598 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370608 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vmwc\" (UniqueName: \"kubernetes.io/projected/cd038612-82db-4064-9515-9f4563121d37-kube-api-access-9vmwc\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370618 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370627 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370635 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd038612-82db-4064-9515-9f4563121d37-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370642 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57e76c2-a16c-43a6-96c0-6b891618e5c0-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.370651 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4w7h\" (UniqueName: \"kubernetes.io/projected/e57e76c2-a16c-43a6-96c0-6b891618e5c0-kube-api-access-w4w7h\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.568502 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b766f8e-6746-48b9-8bc9-3d877bb8cf1b" path="/var/lib/kubelet/pods/7b766f8e-6746-48b9-8bc9-3d877bb8cf1b/volumes" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.569206 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bdbc0d0-444f-495f-bd72-57c41c6bcd71" path="/var/lib/kubelet/pods/8bdbc0d0-444f-495f-bd72-57c41c6bcd71/volumes" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.925353 4492 generic.go:334] "Generic (PLEG): container finished" podID="a789a142-4568-4e21-a16e-75749835da10" containerID="50e35914e8b32ad8e5afee1b89a81666c70528f93d7eb47d3b9d7abbe7eddc95" exitCode=1 Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.925448 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d6f59fdc-lr7g6" event={"ID":"a789a142-4568-4e21-a16e-75749835da10","Type":"ContainerDied","Data":"50e35914e8b32ad8e5afee1b89a81666c70528f93d7eb47d3b9d7abbe7eddc95"} Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.925911 4492 scope.go:117] "RemoveContainer" containerID="27fb0a610089d7a59cca6ac577c161ee12efa632b8c1120837bb874837d1d5ea" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.926513 4492 scope.go:117] "RemoveContainer" containerID="50e35914e8b32ad8e5afee1b89a81666c70528f93d7eb47d3b9d7abbe7eddc95" Feb 20 06:58:05 crc kubenswrapper[4492]: E0220 06:58:05.926902 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5d6f59fdc-lr7g6_openstack(a789a142-4568-4e21-a16e-75749835da10)\"" pod="openstack/heat-api-5d6f59fdc-lr7g6" podUID="a789a142-4568-4e21-a16e-75749835da10" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.928548 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" event={"ID":"cd038612-82db-4064-9515-9f4563121d37","Type":"ContainerDied","Data":"ccfba3bac889921de8993172e3c8abdf1aaac7094cbc07e106914323ecd28945"} Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.928633 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-55d4487b7d-v89bg" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.933825 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerStarted","Data":"2af4ea569c060236996c060b987f04334d954d8f83d952f8cef1789eb2f6bde3"} Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.935596 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6cd457fdb6-qlxrk" event={"ID":"e57e76c2-a16c-43a6-96c0-6b891618e5c0","Type":"ContainerDied","Data":"2a65f27167d6eb2c38e2c78cbd55261156bb30c414aad9b7bd9f1d9c2474f4b0"} Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.935612 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6cd457fdb6-qlxrk" Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.940928 4492 generic.go:334] "Generic (PLEG): container finished" podID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerID="54255656a9f87e93a0f3544332468a4c798375ad113bf77fc87cef02cd4e1c63" exitCode=1 Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.940962 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c46948df-55k9j" event={"ID":"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392","Type":"ContainerDied","Data":"54255656a9f87e93a0f3544332468a4c798375ad113bf77fc87cef02cd4e1c63"} Feb 20 06:58:05 crc kubenswrapper[4492]: I0220 06:58:05.941503 4492 scope.go:117] "RemoveContainer" containerID="54255656a9f87e93a0f3544332468a4c798375ad113bf77fc87cef02cd4e1c63" Feb 20 06:58:05 crc kubenswrapper[4492]: E0220 06:58:05.941761 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-8c46948df-55k9j_openstack(5bddc2a7-62bf-41a1-8bc4-7aeb6c913392)\"" pod="openstack/heat-cfnapi-8c46948df-55k9j" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" Feb 20 06:58:05 crc kubenswrapper[4492]: W0220 06:58:05.987420 4492 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b766f8e_6746_48b9_8bc9_3d877bb8cf1b.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b766f8e_6746_48b9_8bc9_3d877bb8cf1b.slice: no such file or directory Feb 20 06:58:05 crc kubenswrapper[4492]: W0220 06:58:05.988601 4492 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd038612_82db_4064_9515_9f4563121d37.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd038612_82db_4064_9515_9f4563121d37.slice: no such file or directory Feb 20 06:58:05 crc kubenswrapper[4492]: W0220 06:58:05.988960 4492 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode57e76c2_a16c_43a6_96c0_6b891618e5c0.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode57e76c2_a16c_43a6_96c0_6b891618e5c0.slice: no such file or directory Feb 20 06:58:06 crc kubenswrapper[4492]: W0220 06:58:06.009509 4492 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62194cf_bf99_4a79_9d4e_fa7a0f88b8a7.slice/crio-8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac": error while statting cgroup v2: [read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62194cf_bf99_4a79_9d4e_fa7a0f88b8a7.slice/crio-8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac/memory.current: no such device], continuing to push stats Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.028811 4492 scope.go:117] "RemoveContainer" containerID="d6a493a97cb0c02bb235a48fc228b9bc57e346d36c7210589dd10e7d2c3f7c20" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.088779 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-55d4487b7d-v89bg"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.130241 4492 scope.go:117] "RemoveContainer" containerID="20fbefa0675a7b6dcc5cb14869b50309451cb5a576d8e7d85a582d7349fc6482" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.156286 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-55d4487b7d-v89bg"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.209920 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6cd457fdb6-qlxrk"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.248752 4492 scope.go:117] "RemoveContainer" containerID="4e1a85ff90fbd67d56f4fb7e4f1f68b12f035e4799df9cca07f88a3425f7a941" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.256200 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6cd457fdb6-qlxrk"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.272468 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-6tdxw"] Feb 20 06:58:06 crc kubenswrapper[4492]: E0220 06:58:06.272944 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57e76c2-a16c-43a6-96c0-6b891618e5c0" containerName="heat-api" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.272960 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57e76c2-a16c-43a6-96c0-6b891618e5c0" containerName="heat-api" Feb 20 06:58:06 crc kubenswrapper[4492]: E0220 06:58:06.272982 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd038612-82db-4064-9515-9f4563121d37" containerName="heat-cfnapi" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.272988 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd038612-82db-4064-9515-9f4563121d37" containerName="heat-cfnapi" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.273192 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57e76c2-a16c-43a6-96c0-6b891618e5c0" containerName="heat-api" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.273208 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd038612-82db-4064-9515-9f4563121d37" containerName="heat-cfnapi" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.273931 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.315546 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6tdxw"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.351819 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jwgdt"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.352904 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.386558 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jwgdt"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.396220 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm497\" (UniqueName: \"kubernetes.io/projected/33867d4d-53f0-4f25-b61b-f02a98c3b047-kube-api-access-xm497\") pod \"nova-api-db-create-6tdxw\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.404172 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33867d4d-53f0-4f25-b61b-f02a98c3b047-operator-scripts\") pod \"nova-api-db-create-6tdxw\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.404324 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e96057e9-436d-4a08-bf9d-6ae2de326502-operator-scripts\") pod \"nova-cell0-db-create-jwgdt\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.404633 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5ww7\" (UniqueName: \"kubernetes.io/projected/e96057e9-436d-4a08-bf9d-6ae2de326502-kube-api-access-h5ww7\") pod \"nova-cell0-db-create-jwgdt\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.427531 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-be0c-account-create-update-ss6pn"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.429074 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.432918 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.435435 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-be0c-account-create-update-ss6pn"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.442525 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-g5r5h"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.449350 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.461323 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g5r5h"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.534612 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8xbp\" (UniqueName: \"kubernetes.io/projected/f88957ee-9211-4a0e-98f0-0a609cbe03bb-kube-api-access-m8xbp\") pod \"nova-api-be0c-account-create-update-ss6pn\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.534675 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm497\" (UniqueName: \"kubernetes.io/projected/33867d4d-53f0-4f25-b61b-f02a98c3b047-kube-api-access-xm497\") pod \"nova-api-db-create-6tdxw\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.534837 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f88957ee-9211-4a0e-98f0-0a609cbe03bb-operator-scripts\") pod \"nova-api-be0c-account-create-update-ss6pn\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.534924 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33867d4d-53f0-4f25-b61b-f02a98c3b047-operator-scripts\") pod \"nova-api-db-create-6tdxw\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.534975 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e96057e9-436d-4a08-bf9d-6ae2de326502-operator-scripts\") pod \"nova-cell0-db-create-jwgdt\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.535305 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwdnl\" (UniqueName: \"kubernetes.io/projected/f6b92062-3c5c-40e0-9ccf-72e9f896a785-kube-api-access-cwdnl\") pod \"nova-cell1-db-create-g5r5h\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.535618 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6b92062-3c5c-40e0-9ccf-72e9f896a785-operator-scripts\") pod \"nova-cell1-db-create-g5r5h\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.535644 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5ww7\" (UniqueName: \"kubernetes.io/projected/e96057e9-436d-4a08-bf9d-6ae2de326502-kube-api-access-h5ww7\") pod \"nova-cell0-db-create-jwgdt\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.536458 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33867d4d-53f0-4f25-b61b-f02a98c3b047-operator-scripts\") pod \"nova-api-db-create-6tdxw\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.537395 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e96057e9-436d-4a08-bf9d-6ae2de326502-operator-scripts\") pod \"nova-cell0-db-create-jwgdt\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: E0220 06:58:06.569397 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bdbc0d0_444f_495f_bd72_57c41c6bcd71.slice/crio-conmon-2ce5643fc51d0db7366057f9d56f5d38c433a025228ec00329e5e4047cc3d2f7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62194cf_bf99_4a79_9d4e_fa7a0f88b8a7.slice/crio-conmon-953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bdbc0d0_444f_495f_bd72_57c41c6bcd71.slice/crio-2ce5643fc51d0db7366057f9d56f5d38c433a025228ec00329e5e4047cc3d2f7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bdbc0d0_444f_495f_bd72_57c41c6bcd71.slice/crio-6592d07ab3200040ce1380a5185adb60d9d213deaf37653b2f5645919867d4a9\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf62194cf_bf99_4a79_9d4e_fa7a0f88b8a7.slice/crio-8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac\": RecentStats: unable to find data in memory cache]" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.569707 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm497\" (UniqueName: \"kubernetes.io/projected/33867d4d-53f0-4f25-b61b-f02a98c3b047-kube-api-access-xm497\") pod \"nova-api-db-create-6tdxw\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.571091 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5ww7\" (UniqueName: \"kubernetes.io/projected/e96057e9-436d-4a08-bf9d-6ae2de326502-kube-api-access-h5ww7\") pod \"nova-cell0-db-create-jwgdt\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.578509 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ea34-account-create-update-f96zd"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.581999 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.588319 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.623359 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ea34-account-create-update-f96zd"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.638716 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8xbp\" (UniqueName: \"kubernetes.io/projected/f88957ee-9211-4a0e-98f0-0a609cbe03bb-kube-api-access-m8xbp\") pod \"nova-api-be0c-account-create-update-ss6pn\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.638810 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f88957ee-9211-4a0e-98f0-0a609cbe03bb-operator-scripts\") pod \"nova-api-be0c-account-create-update-ss6pn\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.638865 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg44j\" (UniqueName: \"kubernetes.io/projected/081589a6-39f6-4bb6-970d-194e8fa6bc42-kube-api-access-cg44j\") pod \"nova-cell0-ea34-account-create-update-f96zd\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.639134 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwdnl\" (UniqueName: \"kubernetes.io/projected/f6b92062-3c5c-40e0-9ccf-72e9f896a785-kube-api-access-cwdnl\") pod \"nova-cell1-db-create-g5r5h\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.639264 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6b92062-3c5c-40e0-9ccf-72e9f896a785-operator-scripts\") pod \"nova-cell1-db-create-g5r5h\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.639490 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081589a6-39f6-4bb6-970d-194e8fa6bc42-operator-scripts\") pod \"nova-cell0-ea34-account-create-update-f96zd\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.640520 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f88957ee-9211-4a0e-98f0-0a609cbe03bb-operator-scripts\") pod \"nova-api-be0c-account-create-update-ss6pn\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.641044 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6b92062-3c5c-40e0-9ccf-72e9f896a785-operator-scripts\") pod \"nova-cell1-db-create-g5r5h\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.642283 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.670914 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8xbp\" (UniqueName: \"kubernetes.io/projected/f88957ee-9211-4a0e-98f0-0a609cbe03bb-kube-api-access-m8xbp\") pod \"nova-api-be0c-account-create-update-ss6pn\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.671579 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.672466 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwdnl\" (UniqueName: \"kubernetes.io/projected/f6b92062-3c5c-40e0-9ccf-72e9f896a785-kube-api-access-cwdnl\") pod \"nova-cell1-db-create-g5r5h\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.718743 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f823-account-create-update-hkhrt"] Feb 20 06:58:06 crc kubenswrapper[4492]: E0220 06:58:06.719185 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon-log" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.719204 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon-log" Feb 20 06:58:06 crc kubenswrapper[4492]: E0220 06:58:06.719241 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.719249 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.719415 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon-log" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.719431 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerName="horizon" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.733830 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f823-account-create-update-hkhrt"] Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.735649 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.737387 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.737872 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.740326 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-scripts\") pod \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.740398 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-combined-ca-bundle\") pod \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.740435 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-logs\") pod \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.740463 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-secret-key\") pod \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.740615 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-tls-certs\") pod \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.740633 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zhb8\" (UniqueName: \"kubernetes.io/projected/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-kube-api-access-8zhb8\") pod \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.740769 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-config-data\") pod \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\" (UID: \"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7\") " Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.741049 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081589a6-39f6-4bb6-970d-194e8fa6bc42-operator-scripts\") pod \"nova-cell0-ea34-account-create-update-f96zd\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.741152 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg44j\" (UniqueName: \"kubernetes.io/projected/081589a6-39f6-4bb6-970d-194e8fa6bc42-kube-api-access-cg44j\") pod \"nova-cell0-ea34-account-create-update-f96zd\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.744192 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-logs" (OuterVolumeSpecName: "logs") pod "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" (UID: "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.744668 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081589a6-39f6-4bb6-970d-194e8fa6bc42-operator-scripts\") pod \"nova-cell0-ea34-account-create-update-f96zd\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.746222 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.762458 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-kube-api-access-8zhb8" (OuterVolumeSpecName: "kube-api-access-8zhb8") pod "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" (UID: "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7"). InnerVolumeSpecName "kube-api-access-8zhb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.765125 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" (UID: "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.766418 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg44j\" (UniqueName: \"kubernetes.io/projected/081589a6-39f6-4bb6-970d-194e8fa6bc42-kube-api-access-cg44j\") pod \"nova-cell0-ea34-account-create-update-f96zd\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.787809 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.804265 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-config-data" (OuterVolumeSpecName: "config-data") pod "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" (UID: "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.805610 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-scripts" (OuterVolumeSpecName: "scripts") pod "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" (UID: "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.809177 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" (UID: "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.842637 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n4w9\" (UniqueName: \"kubernetes.io/projected/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-kube-api-access-6n4w9\") pod \"nova-cell1-f823-account-create-update-hkhrt\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.842971 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-operator-scripts\") pod \"nova-cell1-f823-account-create-update-hkhrt\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.843106 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.843117 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.843125 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.843133 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.843141 4492 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.843149 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zhb8\" (UniqueName: \"kubernetes.io/projected/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-kube-api-access-8zhb8\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.852600 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" (UID: "f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.913383 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.944521 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n4w9\" (UniqueName: \"kubernetes.io/projected/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-kube-api-access-6n4w9\") pod \"nova-cell1-f823-account-create-update-hkhrt\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.944596 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-operator-scripts\") pod \"nova-cell1-f823-account-create-update-hkhrt\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.944708 4492 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.945339 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-operator-scripts\") pod \"nova-cell1-f823-account-create-update-hkhrt\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.968418 4492 scope.go:117] "RemoveContainer" containerID="54255656a9f87e93a0f3544332468a4c798375ad113bf77fc87cef02cd4e1c63" Feb 20 06:58:06 crc kubenswrapper[4492]: E0220 06:58:06.968618 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-8c46948df-55k9j_openstack(5bddc2a7-62bf-41a1-8bc4-7aeb6c913392)\"" pod="openstack/heat-cfnapi-8c46948df-55k9j" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.974838 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n4w9\" (UniqueName: \"kubernetes.io/projected/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-kube-api-access-6n4w9\") pod \"nova-cell1-f823-account-create-update-hkhrt\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:06 crc kubenswrapper[4492]: I0220 06:58:06.980116 4492 scope.go:117] "RemoveContainer" containerID="50e35914e8b32ad8e5afee1b89a81666c70528f93d7eb47d3b9d7abbe7eddc95" Feb 20 06:58:06 crc kubenswrapper[4492]: E0220 06:58:06.980283 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5d6f59fdc-lr7g6_openstack(a789a142-4568-4e21-a16e-75749835da10)\"" pod="openstack/heat-api-5d6f59fdc-lr7g6" podUID="a789a142-4568-4e21-a16e-75749835da10" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.013419 4492 generic.go:334] "Generic (PLEG): container finished" podID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" containerID="953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba" exitCode=137 Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.013493 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65f5db898d-bxntp" event={"ID":"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7","Type":"ContainerDied","Data":"953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba"} Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.013520 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-65f5db898d-bxntp" event={"ID":"f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7","Type":"ContainerDied","Data":"8bfd009f78a221c6f4a145a2368891bf9c3dd2e3cc804cb355f559f2529e02ac"} Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.013538 4492 scope.go:117] "RemoveContainer" containerID="dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.013628 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-65f5db898d-bxntp" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.037831 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerStarted","Data":"5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58"} Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.096086 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-65f5db898d-bxntp"] Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.150233 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-65f5db898d-bxntp"] Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.260531 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.526513 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6tdxw"] Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.569177 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd038612-82db-4064-9515-9f4563121d37" path="/var/lib/kubelet/pods/cd038612-82db-4064-9515-9f4563121d37/volumes" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.569823 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e57e76c2-a16c-43a6-96c0-6b891618e5c0" path="/var/lib/kubelet/pods/e57e76c2-a16c-43a6-96c0-6b891618e5c0/volumes" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.573461 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7" path="/var/lib/kubelet/pods/f62194cf-bf99-4a79-9d4e-fa7a0f88b8a7/volumes" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.624285 4492 scope.go:117] "RemoveContainer" containerID="953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba" Feb 20 06:58:07 crc kubenswrapper[4492]: W0220 06:58:07.629722 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33867d4d_53f0_4f25_b61b_f02a98c3b047.slice/crio-f6137ae7fac20d713e168705e58708d06f9ea26a5945781c4c36c14ed7fd4aa2 WatchSource:0}: Error finding container f6137ae7fac20d713e168705e58708d06f9ea26a5945781c4c36c14ed7fd4aa2: Status 404 returned error can't find the container with id f6137ae7fac20d713e168705e58708d06f9ea26a5945781c4c36c14ed7fd4aa2 Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.663820 4492 scope.go:117] "RemoveContainer" containerID="dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec" Feb 20 06:58:07 crc kubenswrapper[4492]: E0220 06:58:07.670946 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec\": container with ID starting with dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec not found: ID does not exist" containerID="dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.670982 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec"} err="failed to get container status \"dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec\": rpc error: code = NotFound desc = could not find container \"dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec\": container with ID starting with dfe842505102a0bbbf18d8d3b86adeeb406cfe213294dfc124394e4ecaa4e9ec not found: ID does not exist" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.671003 4492 scope.go:117] "RemoveContainer" containerID="953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba" Feb 20 06:58:07 crc kubenswrapper[4492]: E0220 06:58:07.680727 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba\": container with ID starting with 953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba not found: ID does not exist" containerID="953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.680758 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba"} err="failed to get container status \"953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba\": rpc error: code = NotFound desc = could not find container \"953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba\": container with ID starting with 953e1cbf215763486228a8fd167940ae6904e23b1613e9262ff6e88cf45858ba not found: ID does not exist" Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.745939 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-be0c-account-create-update-ss6pn"] Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.780382 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-g5r5h"] Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.790277 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jwgdt"] Feb 20 06:58:07 crc kubenswrapper[4492]: W0220 06:58:07.836426 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode96057e9_436d_4a08_bf9d_6ae2de326502.slice/crio-bc8e447a08aec4402b76f359449c3722ce38c56a2bd8d482d42ba459744cccad WatchSource:0}: Error finding container bc8e447a08aec4402b76f359449c3722ce38c56a2bd8d482d42ba459744cccad: Status 404 returned error can't find the container with id bc8e447a08aec4402b76f359449c3722ce38c56a2bd8d482d42ba459744cccad Feb 20 06:58:07 crc kubenswrapper[4492]: I0220 06:58:07.883291 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ea34-account-create-update-f96zd"] Feb 20 06:58:08 crc kubenswrapper[4492]: I0220 06:58:08.072549 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ea34-account-create-update-f96zd" event={"ID":"081589a6-39f6-4bb6-970d-194e8fa6bc42","Type":"ContainerStarted","Data":"e9d6953dda9b39cb64243c701e0ba9bafaaa5345563c0c126836ccaa397357cd"} Feb 20 06:58:08 crc kubenswrapper[4492]: I0220 06:58:08.074558 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerStarted","Data":"05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96"} Feb 20 06:58:08 crc kubenswrapper[4492]: I0220 06:58:08.083989 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g5r5h" event={"ID":"f6b92062-3c5c-40e0-9ccf-72e9f896a785","Type":"ContainerStarted","Data":"4245593b83a5c0e53f7e69dae530f39fa4be5b305b213456e52860d0b33aab28"} Feb 20 06:58:08 crc kubenswrapper[4492]: I0220 06:58:08.094831 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-be0c-account-create-update-ss6pn" event={"ID":"f88957ee-9211-4a0e-98f0-0a609cbe03bb","Type":"ContainerStarted","Data":"577041e661fbf82f26d89f272b32ddf164ce8a705d667f5c0e5fda15c42bcf70"} Feb 20 06:58:08 crc kubenswrapper[4492]: I0220 06:58:08.097591 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jwgdt" event={"ID":"e96057e9-436d-4a08-bf9d-6ae2de326502","Type":"ContainerStarted","Data":"bc8e447a08aec4402b76f359449c3722ce38c56a2bd8d482d42ba459744cccad"} Feb 20 06:58:08 crc kubenswrapper[4492]: I0220 06:58:08.099468 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6tdxw" event={"ID":"33867d4d-53f0-4f25-b61b-f02a98c3b047","Type":"ContainerStarted","Data":"f6137ae7fac20d713e168705e58708d06f9ea26a5945781c4c36c14ed7fd4aa2"} Feb 20 06:58:08 crc kubenswrapper[4492]: I0220 06:58:08.225369 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f823-account-create-update-hkhrt"] Feb 20 06:58:08 crc kubenswrapper[4492]: W0220 06:58:08.257124 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc405394d_8e8e_4c21_ae6e_9de9c630c4c3.slice/crio-b8f3055f130ff770bbcf09c80f0c69e3e149783e7f70e653a2ec063d305d1014 WatchSource:0}: Error finding container b8f3055f130ff770bbcf09c80f0c69e3e149783e7f70e653a2ec063d305d1014: Status 404 returned error can't find the container with id b8f3055f130ff770bbcf09c80f0c69e3e149783e7f70e653a2ec063d305d1014 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.108214 4492 generic.go:334] "Generic (PLEG): container finished" podID="e96057e9-436d-4a08-bf9d-6ae2de326502" containerID="e2b485e3a95d0f7b13fdfe393188e41df1e0ddaadbf4bfa4485f53d2d8f7543f" exitCode=0 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.108266 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jwgdt" event={"ID":"e96057e9-436d-4a08-bf9d-6ae2de326502","Type":"ContainerDied","Data":"e2b485e3a95d0f7b13fdfe393188e41df1e0ddaadbf4bfa4485f53d2d8f7543f"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.111419 4492 generic.go:334] "Generic (PLEG): container finished" podID="33867d4d-53f0-4f25-b61b-f02a98c3b047" containerID="42d5e00fa14a4d8a12fcabf40684c00c84bd593c160f043fbcd340afa3f4128c" exitCode=0 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.111488 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6tdxw" event={"ID":"33867d4d-53f0-4f25-b61b-f02a98c3b047","Type":"ContainerDied","Data":"42d5e00fa14a4d8a12fcabf40684c00c84bd593c160f043fbcd340afa3f4128c"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.116724 4492 generic.go:334] "Generic (PLEG): container finished" podID="c405394d-8e8e-4c21-ae6e-9de9c630c4c3" containerID="7c135289b2f6d8cf45477101b9b80c6d3e0e29428418687222ef99ee34b7de3f" exitCode=0 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.116778 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f823-account-create-update-hkhrt" event={"ID":"c405394d-8e8e-4c21-ae6e-9de9c630c4c3","Type":"ContainerDied","Data":"7c135289b2f6d8cf45477101b9b80c6d3e0e29428418687222ef99ee34b7de3f"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.116804 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f823-account-create-update-hkhrt" event={"ID":"c405394d-8e8e-4c21-ae6e-9de9c630c4c3","Type":"ContainerStarted","Data":"b8f3055f130ff770bbcf09c80f0c69e3e149783e7f70e653a2ec063d305d1014"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.122192 4492 generic.go:334] "Generic (PLEG): container finished" podID="081589a6-39f6-4bb6-970d-194e8fa6bc42" containerID="bcfbae3e901b7ae2b74b583bb82a015b944ad3450116bc17d5528d80d396cea3" exitCode=0 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.122264 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ea34-account-create-update-f96zd" event={"ID":"081589a6-39f6-4bb6-970d-194e8fa6bc42","Type":"ContainerDied","Data":"bcfbae3e901b7ae2b74b583bb82a015b944ad3450116bc17d5528d80d396cea3"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.125588 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerStarted","Data":"f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.140979 4492 generic.go:334] "Generic (PLEG): container finished" podID="f6b92062-3c5c-40e0-9ccf-72e9f896a785" containerID="897fe67f3c9ef560a51c7c60257af140e5abb4878b990c0331400b8211873e6d" exitCode=0 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.141047 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g5r5h" event={"ID":"f6b92062-3c5c-40e0-9ccf-72e9f896a785","Type":"ContainerDied","Data":"897fe67f3c9ef560a51c7c60257af140e5abb4878b990c0331400b8211873e6d"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.147863 4492 generic.go:334] "Generic (PLEG): container finished" podID="f88957ee-9211-4a0e-98f0-0a609cbe03bb" containerID="c217337af4e2aca3a2b02e87f11936f17070bb9e635a3363dfcce2813ce50543" exitCode=0 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.147928 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-be0c-account-create-update-ss6pn" event={"ID":"f88957ee-9211-4a0e-98f0-0a609cbe03bb","Type":"ContainerDied","Data":"c217337af4e2aca3a2b02e87f11936f17070bb9e635a3363dfcce2813ce50543"} Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.163861 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.164046 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-log" containerID="cri-o://d383ad6877a1cecb15de4177e5a8f4e75cae890d3e67c1c0fecf31b099218869" gracePeriod=30 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.164156 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-httpd" containerID="cri-o://fbe8afa1ed5c0e0186d700ca9b8ac3b53341dc72899bf4679af6c90d025a3f82" gracePeriod=30 Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.220192 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.221552 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.222197 4492 scope.go:117] "RemoveContainer" containerID="50e35914e8b32ad8e5afee1b89a81666c70528f93d7eb47d3b9d7abbe7eddc95" Feb 20 06:58:09 crc kubenswrapper[4492]: E0220 06:58:09.222585 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5d6f59fdc-lr7g6_openstack(a789a142-4568-4e21-a16e-75749835da10)\"" pod="openstack/heat-api-5d6f59fdc-lr7g6" podUID="a789a142-4568-4e21-a16e-75749835da10" Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.251967 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.252738 4492 scope.go:117] "RemoveContainer" containerID="54255656a9f87e93a0f3544332468a4c798375ad113bf77fc87cef02cd4e1c63" Feb 20 06:58:09 crc kubenswrapper[4492]: E0220 06:58:09.252944 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-8c46948df-55k9j_openstack(5bddc2a7-62bf-41a1-8bc4-7aeb6c913392)\"" pod="openstack/heat-cfnapi-8c46948df-55k9j" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" Feb 20 06:58:09 crc kubenswrapper[4492]: I0220 06:58:09.253054 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.175080 4492 generic.go:334] "Generic (PLEG): container finished" podID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerID="d383ad6877a1cecb15de4177e5a8f4e75cae890d3e67c1c0fecf31b099218869" exitCode=143 Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.175318 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b76c1c3-dda9-4648-a623-83d73d2168dc","Type":"ContainerDied","Data":"d383ad6877a1cecb15de4177e5a8f4e75cae890d3e67c1c0fecf31b099218869"} Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.176180 4492 scope.go:117] "RemoveContainer" containerID="54255656a9f87e93a0f3544332468a4c798375ad113bf77fc87cef02cd4e1c63" Feb 20 06:58:10 crc kubenswrapper[4492]: E0220 06:58:10.176539 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-8c46948df-55k9j_openstack(5bddc2a7-62bf-41a1-8bc4-7aeb6c913392)\"" pod="openstack/heat-cfnapi-8c46948df-55k9j" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.176846 4492 scope.go:117] "RemoveContainer" containerID="50e35914e8b32ad8e5afee1b89a81666c70528f93d7eb47d3b9d7abbe7eddc95" Feb 20 06:58:10 crc kubenswrapper[4492]: E0220 06:58:10.177169 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-5d6f59fdc-lr7g6_openstack(a789a142-4568-4e21-a16e-75749835da10)\"" pod="openstack/heat-api-5d6f59fdc-lr7g6" podUID="a789a142-4568-4e21-a16e-75749835da10" Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.179961 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.180195 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-log" containerID="cri-o://273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747" gracePeriod=30 Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.180694 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-httpd" containerID="cri-o://6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08" gracePeriod=30 Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.847589 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.976854 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6b92062-3c5c-40e0-9ccf-72e9f896a785-operator-scripts\") pod \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.976899 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwdnl\" (UniqueName: \"kubernetes.io/projected/f6b92062-3c5c-40e0-9ccf-72e9f896a785-kube-api-access-cwdnl\") pod \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\" (UID: \"f6b92062-3c5c-40e0-9ccf-72e9f896a785\") " Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.977830 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b92062-3c5c-40e0-9ccf-72e9f896a785-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f6b92062-3c5c-40e0-9ccf-72e9f896a785" (UID: "f6b92062-3c5c-40e0-9ccf-72e9f896a785"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.984650 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b92062-3c5c-40e0-9ccf-72e9f896a785-kube-api-access-cwdnl" (OuterVolumeSpecName: "kube-api-access-cwdnl") pod "f6b92062-3c5c-40e0-9ccf-72e9f896a785" (UID: "f6b92062-3c5c-40e0-9ccf-72e9f896a785"). InnerVolumeSpecName "kube-api-access-cwdnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:10 crc kubenswrapper[4492]: I0220 06:58:10.995193 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.026270 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.050894 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.057895 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.079069 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33867d4d-53f0-4f25-b61b-f02a98c3b047-operator-scripts\") pod \"33867d4d-53f0-4f25-b61b-f02a98c3b047\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.079136 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-operator-scripts\") pod \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.079263 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm497\" (UniqueName: \"kubernetes.io/projected/33867d4d-53f0-4f25-b61b-f02a98c3b047-kube-api-access-xm497\") pod \"33867d4d-53f0-4f25-b61b-f02a98c3b047\" (UID: \"33867d4d-53f0-4f25-b61b-f02a98c3b047\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.079495 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n4w9\" (UniqueName: \"kubernetes.io/projected/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-kube-api-access-6n4w9\") pod \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\" (UID: \"c405394d-8e8e-4c21-ae6e-9de9c630c4c3\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.079901 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c405394d-8e8e-4c21-ae6e-9de9c630c4c3" (UID: "c405394d-8e8e-4c21-ae6e-9de9c630c4c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.081065 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33867d4d-53f0-4f25-b61b-f02a98c3b047-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "33867d4d-53f0-4f25-b61b-f02a98c3b047" (UID: "33867d4d-53f0-4f25-b61b-f02a98c3b047"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.083095 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33867d4d-53f0-4f25-b61b-f02a98c3b047-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.083118 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.083128 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6b92062-3c5c-40e0-9ccf-72e9f896a785-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.083137 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwdnl\" (UniqueName: \"kubernetes.io/projected/f6b92062-3c5c-40e0-9ccf-72e9f896a785-kube-api-access-cwdnl\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.083313 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33867d4d-53f0-4f25-b61b-f02a98c3b047-kube-api-access-xm497" (OuterVolumeSpecName: "kube-api-access-xm497") pod "33867d4d-53f0-4f25-b61b-f02a98c3b047" (UID: "33867d4d-53f0-4f25-b61b-f02a98c3b047"). InnerVolumeSpecName "kube-api-access-xm497". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.084282 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.095147 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-kube-api-access-6n4w9" (OuterVolumeSpecName: "kube-api-access-6n4w9") pod "c405394d-8e8e-4c21-ae6e-9de9c630c4c3" (UID: "c405394d-8e8e-4c21-ae6e-9de9c630c4c3"). InnerVolumeSpecName "kube-api-access-6n4w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186270 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e96057e9-436d-4a08-bf9d-6ae2de326502-operator-scripts\") pod \"e96057e9-436d-4a08-bf9d-6ae2de326502\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186360 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg44j\" (UniqueName: \"kubernetes.io/projected/081589a6-39f6-4bb6-970d-194e8fa6bc42-kube-api-access-cg44j\") pod \"081589a6-39f6-4bb6-970d-194e8fa6bc42\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186433 4492 generic.go:334] "Generic (PLEG): container finished" podID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerID="273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747" exitCode=143 Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186504 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a","Type":"ContainerDied","Data":"273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186648 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081589a6-39f6-4bb6-970d-194e8fa6bc42-operator-scripts\") pod \"081589a6-39f6-4bb6-970d-194e8fa6bc42\" (UID: \"081589a6-39f6-4bb6-970d-194e8fa6bc42\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186685 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5ww7\" (UniqueName: \"kubernetes.io/projected/e96057e9-436d-4a08-bf9d-6ae2de326502-kube-api-access-h5ww7\") pod \"e96057e9-436d-4a08-bf9d-6ae2de326502\" (UID: \"e96057e9-436d-4a08-bf9d-6ae2de326502\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186709 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8xbp\" (UniqueName: \"kubernetes.io/projected/f88957ee-9211-4a0e-98f0-0a609cbe03bb-kube-api-access-m8xbp\") pod \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186758 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f88957ee-9211-4a0e-98f0-0a609cbe03bb-operator-scripts\") pod \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\" (UID: \"f88957ee-9211-4a0e-98f0-0a609cbe03bb\") " Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.186855 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e96057e9-436d-4a08-bf9d-6ae2de326502-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e96057e9-436d-4a08-bf9d-6ae2de326502" (UID: "e96057e9-436d-4a08-bf9d-6ae2de326502"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.187182 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/081589a6-39f6-4bb6-970d-194e8fa6bc42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "081589a6-39f6-4bb6-970d-194e8fa6bc42" (UID: "081589a6-39f6-4bb6-970d-194e8fa6bc42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.187566 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f88957ee-9211-4a0e-98f0-0a609cbe03bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f88957ee-9211-4a0e-98f0-0a609cbe03bb" (UID: "f88957ee-9211-4a0e-98f0-0a609cbe03bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.187740 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f88957ee-9211-4a0e-98f0-0a609cbe03bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.187758 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e96057e9-436d-4a08-bf9d-6ae2de326502-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.187768 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm497\" (UniqueName: \"kubernetes.io/projected/33867d4d-53f0-4f25-b61b-f02a98c3b047-kube-api-access-xm497\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.187778 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n4w9\" (UniqueName: \"kubernetes.io/projected/c405394d-8e8e-4c21-ae6e-9de9c630c4c3-kube-api-access-6n4w9\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.187785 4492 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081589a6-39f6-4bb6-970d-194e8fa6bc42-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.188941 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f823-account-create-update-hkhrt" event={"ID":"c405394d-8e8e-4c21-ae6e-9de9c630c4c3","Type":"ContainerDied","Data":"b8f3055f130ff770bbcf09c80f0c69e3e149783e7f70e653a2ec063d305d1014"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.188990 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8f3055f130ff770bbcf09c80f0c69e3e149783e7f70e653a2ec063d305d1014" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.189059 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f823-account-create-update-hkhrt" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.190868 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88957ee-9211-4a0e-98f0-0a609cbe03bb-kube-api-access-m8xbp" (OuterVolumeSpecName: "kube-api-access-m8xbp") pod "f88957ee-9211-4a0e-98f0-0a609cbe03bb" (UID: "f88957ee-9211-4a0e-98f0-0a609cbe03bb"). InnerVolumeSpecName "kube-api-access-m8xbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.196163 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081589a6-39f6-4bb6-970d-194e8fa6bc42-kube-api-access-cg44j" (OuterVolumeSpecName: "kube-api-access-cg44j") pod "081589a6-39f6-4bb6-970d-194e8fa6bc42" (UID: "081589a6-39f6-4bb6-970d-194e8fa6bc42"). InnerVolumeSpecName "kube-api-access-cg44j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.198849 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e96057e9-436d-4a08-bf9d-6ae2de326502-kube-api-access-h5ww7" (OuterVolumeSpecName: "kube-api-access-h5ww7") pod "e96057e9-436d-4a08-bf9d-6ae2de326502" (UID: "e96057e9-436d-4a08-bf9d-6ae2de326502"). InnerVolumeSpecName "kube-api-access-h5ww7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.200561 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ea34-account-create-update-f96zd" event={"ID":"081589a6-39f6-4bb6-970d-194e8fa6bc42","Type":"ContainerDied","Data":"e9d6953dda9b39cb64243c701e0ba9bafaaa5345563c0c126836ccaa397357cd"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.200630 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9d6953dda9b39cb64243c701e0ba9bafaaa5345563c0c126836ccaa397357cd" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.200738 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ea34-account-create-update-f96zd" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.205492 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerStarted","Data":"18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.206821 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.223704 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-g5r5h" event={"ID":"f6b92062-3c5c-40e0-9ccf-72e9f896a785","Type":"ContainerDied","Data":"4245593b83a5c0e53f7e69dae530f39fa4be5b305b213456e52860d0b33aab28"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.223755 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4245593b83a5c0e53f7e69dae530f39fa4be5b305b213456e52860d0b33aab28" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.225465 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-be0c-account-create-update-ss6pn" event={"ID":"f88957ee-9211-4a0e-98f0-0a609cbe03bb","Type":"ContainerDied","Data":"577041e661fbf82f26d89f272b32ddf164ce8a705d667f5c0e5fda15c42bcf70"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.225520 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="577041e661fbf82f26d89f272b32ddf164ce8a705d667f5c0e5fda15c42bcf70" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.225604 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-be0c-account-create-update-ss6pn" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.226126 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-g5r5h" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.246726 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jwgdt" event={"ID":"e96057e9-436d-4a08-bf9d-6ae2de326502","Type":"ContainerDied","Data":"bc8e447a08aec4402b76f359449c3722ce38c56a2bd8d482d42ba459744cccad"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.246770 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc8e447a08aec4402b76f359449c3722ce38c56a2bd8d482d42ba459744cccad" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.246851 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jwgdt" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.251379 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6tdxw" event={"ID":"33867d4d-53f0-4f25-b61b-f02a98c3b047","Type":"ContainerDied","Data":"f6137ae7fac20d713e168705e58708d06f9ea26a5945781c4c36c14ed7fd4aa2"} Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.251427 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6137ae7fac20d713e168705e58708d06f9ea26a5945781c4c36c14ed7fd4aa2" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.251505 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6tdxw" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.256504 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9955272429999997 podStartE2EDuration="8.256490169s" podCreationTimestamp="2026-02-20 06:58:03 +0000 UTC" firstStartedPulling="2026-02-20 06:58:05.044324638 +0000 UTC m=+1041.815613617" lastFinishedPulling="2026-02-20 06:58:10.305287566 +0000 UTC m=+1047.076576543" observedRunningTime="2026-02-20 06:58:11.229695109 +0000 UTC m=+1048.000984087" watchObservedRunningTime="2026-02-20 06:58:11.256490169 +0000 UTC m=+1048.027779146" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.281510 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.291187 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5ww7\" (UniqueName: \"kubernetes.io/projected/e96057e9-436d-4a08-bf9d-6ae2de326502-kube-api-access-h5ww7\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.291223 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8xbp\" (UniqueName: \"kubernetes.io/projected/f88957ee-9211-4a0e-98f0-0a609cbe03bb-kube-api-access-m8xbp\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.291238 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg44j\" (UniqueName: \"kubernetes.io/projected/081589a6-39f6-4bb6-970d-194e8fa6bc42-kube-api-access-cg44j\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:11 crc kubenswrapper[4492]: I0220 06:58:11.409026 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:58:12 crc kubenswrapper[4492]: I0220 06:58:12.667563 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-d44b9d589-wkjn4" Feb 20 06:58:12 crc kubenswrapper[4492]: I0220 06:58:12.750168 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5d6f59fdc-lr7g6"] Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.180971 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-86b9576f68-82n82" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.197776 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.265898 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-8c46948df-55k9j"] Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.306067 4492 generic.go:334] "Generic (PLEG): container finished" podID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerID="fbe8afa1ed5c0e0186d700ca9b8ac3b53341dc72899bf4679af6c90d025a3f82" exitCode=0 Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.306163 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b76c1c3-dda9-4648-a623-83d73d2168dc","Type":"ContainerDied","Data":"fbe8afa1ed5c0e0186d700ca9b8ac3b53341dc72899bf4679af6c90d025a3f82"} Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.322324 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-central-agent" containerID="cri-o://5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58" gracePeriod=30 Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.322704 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5d6f59fdc-lr7g6" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.323327 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5d6f59fdc-lr7g6" event={"ID":"a789a142-4568-4e21-a16e-75749835da10","Type":"ContainerDied","Data":"598d455259a6b8b75540134eb35c7efe5d3f20c671aa5f5c449f3f828a41ec51"} Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.323367 4492 scope.go:117] "RemoveContainer" containerID="50e35914e8b32ad8e5afee1b89a81666c70528f93d7eb47d3b9d7abbe7eddc95" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.323575 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-notification-agent" containerID="cri-o://05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96" gracePeriod=30 Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.323582 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="sg-core" containerID="cri-o://f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca" gracePeriod=30 Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.323751 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="proxy-httpd" containerID="cri-o://18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91" gracePeriod=30 Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.356282 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data-custom\") pod \"a789a142-4568-4e21-a16e-75749835da10\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.356635 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgt4q\" (UniqueName: \"kubernetes.io/projected/a789a142-4568-4e21-a16e-75749835da10-kube-api-access-xgt4q\") pod \"a789a142-4568-4e21-a16e-75749835da10\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.356798 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data\") pod \"a789a142-4568-4e21-a16e-75749835da10\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.356875 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-combined-ca-bundle\") pod \"a789a142-4568-4e21-a16e-75749835da10\" (UID: \"a789a142-4568-4e21-a16e-75749835da10\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.368948 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a789a142-4568-4e21-a16e-75749835da10-kube-api-access-xgt4q" (OuterVolumeSpecName: "kube-api-access-xgt4q") pod "a789a142-4568-4e21-a16e-75749835da10" (UID: "a789a142-4568-4e21-a16e-75749835da10"). InnerVolumeSpecName "kube-api-access-xgt4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.369543 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a789a142-4568-4e21-a16e-75749835da10" (UID: "a789a142-4568-4e21-a16e-75749835da10"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.452592 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a789a142-4568-4e21-a16e-75749835da10" (UID: "a789a142-4568-4e21-a16e-75749835da10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.458570 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data" (OuterVolumeSpecName: "config-data") pod "a789a142-4568-4e21-a16e-75749835da10" (UID: "a789a142-4568-4e21-a16e-75749835da10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.465663 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.465725 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.465742 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a789a142-4568-4e21-a16e-75749835da10-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.465753 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgt4q\" (UniqueName: \"kubernetes.io/projected/a789a142-4568-4e21-a16e-75749835da10-kube-api-access-xgt4q\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.530237 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.669689 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-config-data\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.669912 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5d6f59fdc-lr7g6"] Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.669989 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-scripts\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.670098 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-logs\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.670137 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-public-tls-certs\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.670244 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb9nt\" (UniqueName: \"kubernetes.io/projected/6b76c1c3-dda9-4648-a623-83d73d2168dc-kube-api-access-qb9nt\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.670305 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.670344 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-httpd-run\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.670383 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-combined-ca-bundle\") pod \"6b76c1c3-dda9-4648-a623-83d73d2168dc\" (UID: \"6b76c1c3-dda9-4648-a623-83d73d2168dc\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.671824 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.672067 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-logs" (OuterVolumeSpecName: "logs") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.687861 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b76c1c3-dda9-4648-a623-83d73d2168dc-kube-api-access-qb9nt" (OuterVolumeSpecName: "kube-api-access-qb9nt") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "kube-api-access-qb9nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.689291 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-scripts" (OuterVolumeSpecName: "scripts") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.700691 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.716932 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5d6f59fdc-lr7g6"] Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.744664 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.775184 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.775214 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb9nt\" (UniqueName: \"kubernetes.io/projected/6b76c1c3-dda9-4648-a623-83d73d2168dc-kube-api-access-qb9nt\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.775240 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.775253 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b76c1c3-dda9-4648-a623-83d73d2168dc-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.775262 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.775271 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.854529 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.854681 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.879977 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.880010 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.889535 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-config-data" (OuterVolumeSpecName: "config-data") pod "6b76c1c3-dda9-4648-a623-83d73d2168dc" (UID: "6b76c1c3-dda9-4648-a623-83d73d2168dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.892658 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.981009 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data\") pod \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.981072 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9tmc\" (UniqueName: \"kubernetes.io/projected/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-kube-api-access-n9tmc\") pod \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.981173 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data-custom\") pod \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.981269 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-combined-ca-bundle\") pod \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\" (UID: \"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392\") " Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.981954 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b76c1c3-dda9-4648-a623-83d73d2168dc-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.984320 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-kube-api-access-n9tmc" (OuterVolumeSpecName: "kube-api-access-n9tmc") pod "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" (UID: "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392"). InnerVolumeSpecName "kube-api-access-n9tmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:13 crc kubenswrapper[4492]: I0220 06:58:13.989446 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" (UID: "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.012645 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" (UID: "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.079080 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data" (OuterVolumeSpecName: "config-data") pod "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" (UID: "5bddc2a7-62bf-41a1-8bc4-7aeb6c913392"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.083747 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.083777 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.083853 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9tmc\" (UniqueName: \"kubernetes.io/projected/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-kube-api-access-n9tmc\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.083867 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.147817 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185242 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-logs\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185335 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-combined-ca-bundle\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185364 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-httpd-run\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185406 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5np9\" (UniqueName: \"kubernetes.io/projected/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-kube-api-access-v5np9\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185426 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185499 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-scripts\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185520 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-internal-tls-certs\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185615 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-config-data\") pod \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\" (UID: \"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a\") " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.185770 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-logs" (OuterVolumeSpecName: "logs") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.186156 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.186365 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.202580 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.202601 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-scripts" (OuterVolumeSpecName: "scripts") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.211465 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-kube-api-access-v5np9" (OuterVolumeSpecName: "kube-api-access-v5np9") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "kube-api-access-v5np9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.252973 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.287616 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.287644 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.287656 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5np9\" (UniqueName: \"kubernetes.io/projected/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-kube-api-access-v5np9\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.287693 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.287702 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.307601 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.309168 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.320593 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-config-data" (OuterVolumeSpecName: "config-data") pod "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" (UID: "d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.335459 4492 generic.go:334] "Generic (PLEG): container finished" podID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerID="18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91" exitCode=0 Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.335587 4492 generic.go:334] "Generic (PLEG): container finished" podID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerID="f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca" exitCode=2 Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.335655 4492 generic.go:334] "Generic (PLEG): container finished" podID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerID="05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96" exitCode=0 Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.335757 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerDied","Data":"18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91"} Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.335881 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerDied","Data":"f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca"} Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.335959 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerDied","Data":"05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96"} Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.337483 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6b76c1c3-dda9-4648-a623-83d73d2168dc","Type":"ContainerDied","Data":"606bd7f86e5b76be302aa4240e78cf39a3c2e3423318d2daf2a24c733560ac0d"} Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.337774 4492 scope.go:117] "RemoveContainer" containerID="fbe8afa1ed5c0e0186d700ca9b8ac3b53341dc72899bf4679af6c90d025a3f82" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.337606 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.347626 4492 generic.go:334] "Generic (PLEG): container finished" podID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerID="6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08" exitCode=0 Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.347737 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.348924 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8c46948df-55k9j" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.357115 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a","Type":"ContainerDied","Data":"6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08"} Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.357176 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a","Type":"ContainerDied","Data":"e7a13de06402e79c32a7a10172ee37d601bb02e0106ac7f8801ede6e80e1a8c4"} Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.357195 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8c46948df-55k9j" event={"ID":"5bddc2a7-62bf-41a1-8bc4-7aeb6c913392","Type":"ContainerDied","Data":"03214e31f63648e413d7f5d334a5b4143fcc782b13710dea8f2db5fe13863968"} Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.393432 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.393461 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.401272 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.401542 4492 scope.go:117] "RemoveContainer" containerID="d383ad6877a1cecb15de4177e5a8f4e75cae890d3e67c1c0fecf31b099218869" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.435144 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.451973 4492 scope.go:117] "RemoveContainer" containerID="6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.459825 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.487524 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.500779 4492 scope.go:117] "RemoveContainer" containerID="273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.505611 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.527104 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.527466 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a789a142-4568-4e21-a16e-75749835da10" containerName="heat-api" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.527496 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a789a142-4568-4e21-a16e-75749835da10" containerName="heat-api" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.527507 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-log" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.527513 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-log" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532518 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-httpd" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532542 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-httpd" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532557 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-httpd" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532565 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-httpd" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532573 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c405394d-8e8e-4c21-ae6e-9de9c630c4c3" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532579 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c405394d-8e8e-4c21-ae6e-9de9c630c4c3" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532589 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081589a6-39f6-4bb6-970d-194e8fa6bc42" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532598 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="081589a6-39f6-4bb6-970d-194e8fa6bc42" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532615 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerName="heat-cfnapi" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532621 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerName="heat-cfnapi" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532632 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-log" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532637 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-log" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532652 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a789a142-4568-4e21-a16e-75749835da10" containerName="heat-api" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532660 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a789a142-4568-4e21-a16e-75749835da10" containerName="heat-api" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532669 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96057e9-436d-4a08-bf9d-6ae2de326502" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532674 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96057e9-436d-4a08-bf9d-6ae2de326502" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532697 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f88957ee-9211-4a0e-98f0-0a609cbe03bb" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532703 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f88957ee-9211-4a0e-98f0-0a609cbe03bb" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532710 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b92062-3c5c-40e0-9ccf-72e9f896a785" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532716 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b92062-3c5c-40e0-9ccf-72e9f896a785" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532727 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerName="heat-cfnapi" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532732 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerName="heat-cfnapi" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.532745 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33867d4d-53f0-4f25-b61b-f02a98c3b047" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.532751 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="33867d4d-53f0-4f25-b61b-f02a98c3b047" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533003 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-httpd" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533014 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerName="heat-cfnapi" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533025 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" containerName="heat-cfnapi" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533035 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a789a142-4568-4e21-a16e-75749835da10" containerName="heat-api" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533044 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f88957ee-9211-4a0e-98f0-0a609cbe03bb" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533052 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-httpd" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533062 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a789a142-4568-4e21-a16e-75749835da10" containerName="heat-api" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533067 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b92062-3c5c-40e0-9ccf-72e9f896a785" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533075 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="081589a6-39f6-4bb6-970d-194e8fa6bc42" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533082 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" containerName="glance-log" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533091 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" containerName="glance-log" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533099 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c405394d-8e8e-4c21-ae6e-9de9c630c4c3" containerName="mariadb-account-create-update" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533109 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96057e9-436d-4a08-bf9d-6ae2de326502" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.533117 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="33867d4d-53f0-4f25-b61b-f02a98c3b047" containerName="mariadb-database-create" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.534032 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.552147 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.552371 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.552491 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-jvgvq" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.552600 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.555991 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-8c46948df-55k9j"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.560020 4492 scope.go:117] "RemoveContainer" containerID="6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.560369 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-8c46948df-55k9j"] Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.562461 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08\": container with ID starting with 6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08 not found: ID does not exist" containerID="6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.562588 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08"} err="failed to get container status \"6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08\": rpc error: code = NotFound desc = could not find container \"6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08\": container with ID starting with 6c03977a382290d7411e5d8a9f7502fb073bd4bd8454935cf071918cd6e09d08 not found: ID does not exist" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.562658 4492 scope.go:117] "RemoveContainer" containerID="273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747" Feb 20 06:58:14 crc kubenswrapper[4492]: E0220 06:58:14.569403 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747\": container with ID starting with 273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747 not found: ID does not exist" containerID="273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.569448 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747"} err="failed to get container status \"273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747\": rpc error: code = NotFound desc = could not find container \"273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747\": container with ID starting with 273820caaaa155046042fcfa537fd0b64d54f919dcba975506ec008010fc8747 not found: ID does not exist" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.569492 4492 scope.go:117] "RemoveContainer" containerID="54255656a9f87e93a0f3544332468a4c798375ad113bf77fc87cef02cd4e1c63" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.578220 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.597422 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.598873 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.601249 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.601431 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.605457 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607648 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607696 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607747 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607779 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607807 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-logs\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607819 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lslfx\" (UniqueName: \"kubernetes.io/projected/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-kube-api-access-lslfx\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607840 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.607897 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.711840 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.712436 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.712498 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-logs\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.712520 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lslfx\" (UniqueName: \"kubernetes.io/projected/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-kube-api-access-lslfx\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.712546 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.712643 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.713304 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.713373 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.713895 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.713956 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.716600 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-logs\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.717978 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.718330 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.720351 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.723264 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.749464 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.753069 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lslfx\" (UniqueName: \"kubernetes.io/projected/0f6bd7f5-2c98-402f-b8ab-c72e291265e3-kube-api-access-lslfx\") pod \"glance-default-external-api-0\" (UID: \"0f6bd7f5-2c98-402f-b8ab-c72e291265e3\") " pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.815787 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.816000 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/377d765a-232c-4fe2-a066-6865ccd8d677-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.816033 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.816181 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mjs8\" (UniqueName: \"kubernetes.io/projected/377d765a-232c-4fe2-a066-6865ccd8d677-kube-api-access-9mjs8\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.816360 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-config-data\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.816455 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.816526 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/377d765a-232c-4fe2-a066-6865ccd8d677-logs\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.816574 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-scripts\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.863113 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.927352 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mjs8\" (UniqueName: \"kubernetes.io/projected/377d765a-232c-4fe2-a066-6865ccd8d677-kube-api-access-9mjs8\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.931910 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-config-data\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.932004 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.932040 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/377d765a-232c-4fe2-a066-6865ccd8d677-logs\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.932057 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-scripts\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.932173 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.932198 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/377d765a-232c-4fe2-a066-6865ccd8d677-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.932232 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.932899 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.933797 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/377d765a-232c-4fe2-a066-6865ccd8d677-logs\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.934017 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/377d765a-232c-4fe2-a066-6865ccd8d677-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.945594 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mjs8\" (UniqueName: \"kubernetes.io/projected/377d765a-232c-4fe2-a066-6865ccd8d677-kube-api-access-9mjs8\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.946004 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.948904 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-config-data\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.951247 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-scripts\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:14 crc kubenswrapper[4492]: I0220 06:58:14.953460 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/377d765a-232c-4fe2-a066-6865ccd8d677-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.029720 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"377d765a-232c-4fe2-a066-6865ccd8d677\") " pod="openstack/glance-default-internal-api-0" Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.217036 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.493230 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.573157 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bddc2a7-62bf-41a1-8bc4-7aeb6c913392" path="/var/lib/kubelet/pods/5bddc2a7-62bf-41a1-8bc4-7aeb6c913392/volumes" Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.573758 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b76c1c3-dda9-4648-a623-83d73d2168dc" path="/var/lib/kubelet/pods/6b76c1c3-dda9-4648-a623-83d73d2168dc/volumes" Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.574412 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a789a142-4568-4e21-a16e-75749835da10" path="/var/lib/kubelet/pods/a789a142-4568-4e21-a16e-75749835da10/volumes" Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.579412 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a" path="/var/lib/kubelet/pods/d9c5afbb-3107-47eb-9603-8cc1bb2ebf0a/volumes" Feb 20 06:58:15 crc kubenswrapper[4492]: I0220 06:58:15.848127 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.061788 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.071725 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-sg-core-conf-yaml\") pod \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.071839 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcppk\" (UniqueName: \"kubernetes.io/projected/3f9c0e03-bed9-41e5-86c2-633bcf86632f-kube-api-access-zcppk\") pod \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.071871 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-log-httpd\") pod \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.071887 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-scripts\") pod \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.071907 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-config-data\") pod \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.071965 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-combined-ca-bundle\") pod \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.071985 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-run-httpd\") pod \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\" (UID: \"3f9c0e03-bed9-41e5-86c2-633bcf86632f\") " Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.073645 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3f9c0e03-bed9-41e5-86c2-633bcf86632f" (UID: "3f9c0e03-bed9-41e5-86c2-633bcf86632f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.073877 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3f9c0e03-bed9-41e5-86c2-633bcf86632f" (UID: "3f9c0e03-bed9-41e5-86c2-633bcf86632f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.086913 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-scripts" (OuterVolumeSpecName: "scripts") pod "3f9c0e03-bed9-41e5-86c2-633bcf86632f" (UID: "3f9c0e03-bed9-41e5-86c2-633bcf86632f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.088792 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f9c0e03-bed9-41e5-86c2-633bcf86632f-kube-api-access-zcppk" (OuterVolumeSpecName: "kube-api-access-zcppk") pod "3f9c0e03-bed9-41e5-86c2-633bcf86632f" (UID: "3f9c0e03-bed9-41e5-86c2-633bcf86632f"). InnerVolumeSpecName "kube-api-access-zcppk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.144376 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3f9c0e03-bed9-41e5-86c2-633bcf86632f" (UID: "3f9c0e03-bed9-41e5-86c2-633bcf86632f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.174653 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.174676 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcppk\" (UniqueName: \"kubernetes.io/projected/3f9c0e03-bed9-41e5-86c2-633bcf86632f-kube-api-access-zcppk\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.174706 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.174715 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.174722 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f9c0e03-bed9-41e5-86c2-633bcf86632f-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.234594 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f9c0e03-bed9-41e5-86c2-633bcf86632f" (UID: "3f9c0e03-bed9-41e5-86c2-633bcf86632f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.241678 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-config-data" (OuterVolumeSpecName: "config-data") pod "3f9c0e03-bed9-41e5-86c2-633bcf86632f" (UID: "3f9c0e03-bed9-41e5-86c2-633bcf86632f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.277060 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.277088 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f9c0e03-bed9-41e5-86c2-633bcf86632f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.422783 4492 generic.go:334] "Generic (PLEG): container finished" podID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerID="5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58" exitCode=0 Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.422882 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerDied","Data":"5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58"} Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.422910 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f9c0e03-bed9-41e5-86c2-633bcf86632f","Type":"ContainerDied","Data":"2af4ea569c060236996c060b987f04334d954d8f83d952f8cef1789eb2f6bde3"} Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.422933 4492 scope.go:117] "RemoveContainer" containerID="18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.423096 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.433997 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"377d765a-232c-4fe2-a066-6865ccd8d677","Type":"ContainerStarted","Data":"9ff95ecd3c3fd9ffd52752d66ce81fa0b6e36d5eaf43237db084685c98b00936"} Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.435436 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f6bd7f5-2c98-402f-b8ab-c72e291265e3","Type":"ContainerStarted","Data":"24181f95899cf2b65c7c05f55d5a85405e6abccd3f00d7548b26e9b12bfbac1c"} Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.435461 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f6bd7f5-2c98-402f-b8ab-c72e291265e3","Type":"ContainerStarted","Data":"30cfa333631ed932d1f5a2aac0b7c7d1e5cac7ef81c521c8a9203319cf919cfa"} Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.476830 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.488579 4492 scope.go:117] "RemoveContainer" containerID="f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.489663 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.503563 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.503872 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-central-agent" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.503891 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-central-agent" Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.503913 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="sg-core" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.503920 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="sg-core" Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.503935 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-notification-agent" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.503940 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-notification-agent" Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.503952 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="proxy-httpd" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.503957 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="proxy-httpd" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.504110 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="sg-core" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.504129 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="proxy-httpd" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.504141 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-notification-agent" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.504148 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" containerName="ceilometer-central-agent" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.505683 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.519901 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.520188 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.535395 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.562584 4492 scope.go:117] "RemoveContainer" containerID="05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.586203 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-log-httpd\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.586367 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.586427 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5tzh\" (UniqueName: \"kubernetes.io/projected/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-kube-api-access-t5tzh\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.586453 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-config-data\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.586554 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-run-httpd\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.586623 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.586653 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-scripts\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.689906 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.690141 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5tzh\" (UniqueName: \"kubernetes.io/projected/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-kube-api-access-t5tzh\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.690163 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-config-data\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.690183 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-run-httpd\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.690220 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.690241 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-scripts\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.690270 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-log-httpd\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.690671 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-log-httpd\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.691215 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-run-httpd\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.702838 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-config-data\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.703310 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.707967 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.709818 4492 scope.go:117] "RemoveContainer" containerID="5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.713704 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5tzh\" (UniqueName: \"kubernetes.io/projected/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-kube-api-access-t5tzh\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.728312 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-scripts\") pod \"ceilometer-0\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.822560 4492 scope.go:117] "RemoveContainer" containerID="18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91" Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.834582 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91\": container with ID starting with 18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91 not found: ID does not exist" containerID="18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.834617 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91"} err="failed to get container status \"18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91\": rpc error: code = NotFound desc = could not find container \"18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91\": container with ID starting with 18b896cba61b76baf2e1781c03a8d2c15008131defa66ebee3b25a6de4ba3c91 not found: ID does not exist" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.834639 4492 scope.go:117] "RemoveContainer" containerID="f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.834961 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.835351 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca\": container with ID starting with f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca not found: ID does not exist" containerID="f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.835376 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca"} err="failed to get container status \"f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca\": rpc error: code = NotFound desc = could not find container \"f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca\": container with ID starting with f477544f669bbcd656a585a7ef6449705d5f71f0c56488d5f1565688a0f4d8ca not found: ID does not exist" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.835391 4492 scope.go:117] "RemoveContainer" containerID="05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96" Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.836771 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96\": container with ID starting with 05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96 not found: ID does not exist" containerID="05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.836810 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96"} err="failed to get container status \"05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96\": rpc error: code = NotFound desc = could not find container \"05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96\": container with ID starting with 05f91515bae0b00371a1cdd368d563a5cc8a4becb1489a48951fa216fac04c96 not found: ID does not exist" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.836835 4492 scope.go:117] "RemoveContainer" containerID="5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58" Feb 20 06:58:16 crc kubenswrapper[4492]: E0220 06:58:16.840545 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58\": container with ID starting with 5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58 not found: ID does not exist" containerID="5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58" Feb 20 06:58:16 crc kubenswrapper[4492]: I0220 06:58:16.840579 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58"} err="failed to get container status \"5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58\": rpc error: code = NotFound desc = could not find container \"5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58\": container with ID starting with 5a5ef71a64298e201d6c2ec34bd916fe4f7dc1420b6b1c9c43fe66a9536bcc58 not found: ID does not exist" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.213950 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bsh7r"] Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.229159 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.237120 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.237391 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.237535 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xpp5p" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.268220 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bsh7r"] Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.308773 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-scripts\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.308836 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-config-data\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.308939 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.309200 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29wlr\" (UniqueName: \"kubernetes.io/projected/d97b093e-5584-4662-a976-215ada9ef003-kube-api-access-29wlr\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.411979 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.414763 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29wlr\" (UniqueName: \"kubernetes.io/projected/d97b093e-5584-4662-a976-215ada9ef003-kube-api-access-29wlr\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.415048 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-scripts\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.415112 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-config-data\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.426964 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-scripts\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.428187 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.439077 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-config-data\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.452500 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29wlr\" (UniqueName: \"kubernetes.io/projected/d97b093e-5584-4662-a976-215ada9ef003-kube-api-access-29wlr\") pod \"nova-cell0-conductor-db-sync-bsh7r\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.469285 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"377d765a-232c-4fe2-a066-6865ccd8d677","Type":"ContainerStarted","Data":"a668d35e8183d5574a39369fa3de78ca7b0adf40ca2f0bd4f75723edc16a793a"} Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.576794 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f9c0e03-bed9-41e5-86c2-633bcf86632f" path="/var/lib/kubelet/pods/3f9c0e03-bed9-41e5-86c2-633bcf86632f/volumes" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.597861 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:17 crc kubenswrapper[4492]: I0220 06:58:17.646575 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:18 crc kubenswrapper[4492]: I0220 06:58:18.156239 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bsh7r"] Feb 20 06:58:18 crc kubenswrapper[4492]: I0220 06:58:18.482149 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerStarted","Data":"ed8085cd3de2f8eea4306f8968546d749be441902cafcdf6b3e48cca96dbe040"} Feb 20 06:58:18 crc kubenswrapper[4492]: I0220 06:58:18.483679 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" event={"ID":"d97b093e-5584-4662-a976-215ada9ef003","Type":"ContainerStarted","Data":"320bb74a3f71d9d1a7833b863ea38235da98818e7439f04b1a6ea4fe66e62284"} Feb 20 06:58:18 crc kubenswrapper[4492]: I0220 06:58:18.486623 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"377d765a-232c-4fe2-a066-6865ccd8d677","Type":"ContainerStarted","Data":"e5a0aa2fb6f8e0e21270ce463201ca616a1b06c788ef1d16db7e6206a3bbf000"} Feb 20 06:58:18 crc kubenswrapper[4492]: I0220 06:58:18.488658 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f6bd7f5-2c98-402f-b8ab-c72e291265e3","Type":"ContainerStarted","Data":"f1b6a63e05d6332a456fd9a224dddaebccaf2ead8793575443ddc3b1134a4d73"} Feb 20 06:58:18 crc kubenswrapper[4492]: I0220 06:58:18.516861 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.516834404 podStartE2EDuration="4.516834404s" podCreationTimestamp="2026-02-20 06:58:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:18.509497674 +0000 UTC m=+1055.280786652" watchObservedRunningTime="2026-02-20 06:58:18.516834404 +0000 UTC m=+1055.288123383" Feb 20 06:58:18 crc kubenswrapper[4492]: I0220 06:58:18.547679 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.547656845 podStartE2EDuration="4.547656845s" podCreationTimestamp="2026-02-20 06:58:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:18.538293815 +0000 UTC m=+1055.309582794" watchObservedRunningTime="2026-02-20 06:58:18.547656845 +0000 UTC m=+1055.318945823" Feb 20 06:58:19 crc kubenswrapper[4492]: I0220 06:58:19.224273 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-749c7585d6-zr8k4" Feb 20 06:58:19 crc kubenswrapper[4492]: I0220 06:58:19.339003 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-58b9894566-r5jqg"] Feb 20 06:58:19 crc kubenswrapper[4492]: I0220 06:58:19.339191 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-58b9894566-r5jqg" podUID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" containerName="heat-engine" containerID="cri-o://8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" gracePeriod=60 Feb 20 06:58:19 crc kubenswrapper[4492]: I0220 06:58:19.524225 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerStarted","Data":"2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b"} Feb 20 06:58:20 crc kubenswrapper[4492]: I0220 06:58:20.538460 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerStarted","Data":"6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c"} Feb 20 06:58:20 crc kubenswrapper[4492]: I0220 06:58:20.765952 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:21 crc kubenswrapper[4492]: E0220 06:58:21.337041 4492 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 20 06:58:21 crc kubenswrapper[4492]: E0220 06:58:21.338635 4492 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 20 06:58:21 crc kubenswrapper[4492]: E0220 06:58:21.342267 4492 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 20 06:58:21 crc kubenswrapper[4492]: E0220 06:58:21.342304 4492 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-58b9894566-r5jqg" podUID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" containerName="heat-engine" Feb 20 06:58:21 crc kubenswrapper[4492]: I0220 06:58:21.552091 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerStarted","Data":"7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c"} Feb 20 06:58:22 crc kubenswrapper[4492]: I0220 06:58:22.579866 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerStarted","Data":"cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3"} Feb 20 06:58:22 crc kubenswrapper[4492]: I0220 06:58:22.583962 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 06:58:22 crc kubenswrapper[4492]: I0220 06:58:22.580007 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-central-agent" containerID="cri-o://2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b" gracePeriod=30 Feb 20 06:58:22 crc kubenswrapper[4492]: I0220 06:58:22.580061 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-notification-agent" containerID="cri-o://6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c" gracePeriod=30 Feb 20 06:58:22 crc kubenswrapper[4492]: I0220 06:58:22.580090 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="proxy-httpd" containerID="cri-o://cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3" gracePeriod=30 Feb 20 06:58:22 crc kubenswrapper[4492]: I0220 06:58:22.580045 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="sg-core" containerID="cri-o://7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c" gracePeriod=30 Feb 20 06:58:22 crc kubenswrapper[4492]: I0220 06:58:22.605742 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.648096406 podStartE2EDuration="6.605722839s" podCreationTimestamp="2026-02-20 06:58:16 +0000 UTC" firstStartedPulling="2026-02-20 06:58:17.665251006 +0000 UTC m=+1054.436539984" lastFinishedPulling="2026-02-20 06:58:21.622877439 +0000 UTC m=+1058.394166417" observedRunningTime="2026-02-20 06:58:22.600913504 +0000 UTC m=+1059.372202482" watchObservedRunningTime="2026-02-20 06:58:22.605722839 +0000 UTC m=+1059.377011817" Feb 20 06:58:23 crc kubenswrapper[4492]: I0220 06:58:23.600092 4492 generic.go:334] "Generic (PLEG): container finished" podID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerID="cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3" exitCode=0 Feb 20 06:58:23 crc kubenswrapper[4492]: I0220 06:58:23.600122 4492 generic.go:334] "Generic (PLEG): container finished" podID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerID="7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c" exitCode=2 Feb 20 06:58:23 crc kubenswrapper[4492]: I0220 06:58:23.600130 4492 generic.go:334] "Generic (PLEG): container finished" podID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerID="6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c" exitCode=0 Feb 20 06:58:23 crc kubenswrapper[4492]: I0220 06:58:23.600151 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerDied","Data":"cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3"} Feb 20 06:58:23 crc kubenswrapper[4492]: I0220 06:58:23.600203 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerDied","Data":"7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c"} Feb 20 06:58:23 crc kubenswrapper[4492]: I0220 06:58:23.600214 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerDied","Data":"6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c"} Feb 20 06:58:24 crc kubenswrapper[4492]: I0220 06:58:24.877053 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 06:58:24 crc kubenswrapper[4492]: I0220 06:58:24.878887 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 20 06:58:24 crc kubenswrapper[4492]: I0220 06:58:24.947832 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 06:58:24 crc kubenswrapper[4492]: I0220 06:58:24.948352 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.219179 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.219381 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.272415 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.297681 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.646832 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.646983 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.647001 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:25 crc kubenswrapper[4492]: I0220 06:58:25.647012 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 20 06:58:27 crc kubenswrapper[4492]: I0220 06:58:27.661141 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:58:27 crc kubenswrapper[4492]: I0220 06:58:27.661759 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:58:27 crc kubenswrapper[4492]: I0220 06:58:27.661714 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:58:27 crc kubenswrapper[4492]: I0220 06:58:27.661896 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:58:30 crc kubenswrapper[4492]: I0220 06:58:30.368978 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 06:58:30 crc kubenswrapper[4492]: I0220 06:58:30.370302 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:58:30 crc kubenswrapper[4492]: I0220 06:58:30.377408 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 20 06:58:30 crc kubenswrapper[4492]: I0220 06:58:30.554031 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:30 crc kubenswrapper[4492]: I0220 06:58:30.554463 4492 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 20 06:58:30 crc kubenswrapper[4492]: I0220 06:58:30.815436 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 20 06:58:31 crc kubenswrapper[4492]: E0220 06:58:31.336940 4492 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 20 06:58:31 crc kubenswrapper[4492]: E0220 06:58:31.344717 4492 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 20 06:58:31 crc kubenswrapper[4492]: E0220 06:58:31.348760 4492 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Feb 20 06:58:31 crc kubenswrapper[4492]: E0220 06:58:31.348802 4492 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-58b9894566-r5jqg" podUID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" containerName="heat-engine" Feb 20 06:58:31 crc kubenswrapper[4492]: I0220 06:58:31.730546 4492 generic.go:334] "Generic (PLEG): container finished" podID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" exitCode=0 Feb 20 06:58:31 crc kubenswrapper[4492]: I0220 06:58:31.730599 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-58b9894566-r5jqg" event={"ID":"e7da01e6-bb65-47dd-8357-72f1fec9d95d","Type":"ContainerDied","Data":"8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e"} Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.759751 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-58b9894566-r5jqg" event={"ID":"e7da01e6-bb65-47dd-8357-72f1fec9d95d","Type":"ContainerDied","Data":"68432b00028a0d500ea5e2fb5eb7e74a949405f5c23055ddd51370e876bc2db1"} Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.760143 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68432b00028a0d500ea5e2fb5eb7e74a949405f5c23055ddd51370e876bc2db1" Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.821188 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.928607 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-combined-ca-bundle\") pod \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.928691 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data\") pod \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.928791 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7grtm\" (UniqueName: \"kubernetes.io/projected/e7da01e6-bb65-47dd-8357-72f1fec9d95d-kube-api-access-7grtm\") pod \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.929015 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data-custom\") pod \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\" (UID: \"e7da01e6-bb65-47dd-8357-72f1fec9d95d\") " Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.939059 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7da01e6-bb65-47dd-8357-72f1fec9d95d-kube-api-access-7grtm" (OuterVolumeSpecName: "kube-api-access-7grtm") pod "e7da01e6-bb65-47dd-8357-72f1fec9d95d" (UID: "e7da01e6-bb65-47dd-8357-72f1fec9d95d"). InnerVolumeSpecName "kube-api-access-7grtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.939200 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e7da01e6-bb65-47dd-8357-72f1fec9d95d" (UID: "e7da01e6-bb65-47dd-8357-72f1fec9d95d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.954803 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7da01e6-bb65-47dd-8357-72f1fec9d95d" (UID: "e7da01e6-bb65-47dd-8357-72f1fec9d95d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:33 crc kubenswrapper[4492]: I0220 06:58:33.979588 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data" (OuterVolumeSpecName: "config-data") pod "e7da01e6-bb65-47dd-8357-72f1fec9d95d" (UID: "e7da01e6-bb65-47dd-8357-72f1fec9d95d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.030904 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.031121 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.031131 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7grtm\" (UniqueName: \"kubernetes.io/projected/e7da01e6-bb65-47dd-8357-72f1fec9d95d-kube-api-access-7grtm\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.031142 4492 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7da01e6-bb65-47dd-8357-72f1fec9d95d-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.780859 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-58b9894566-r5jqg" Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.781327 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" event={"ID":"d97b093e-5584-4662-a976-215ada9ef003","Type":"ContainerStarted","Data":"7f0f8e79841d1114a5bd817e65c94bd327be8a4c19392bc146b97a3c66ec3466"} Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.804734 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" podStartSLOduration=2.290250905 podStartE2EDuration="17.804716974s" podCreationTimestamp="2026-02-20 06:58:17 +0000 UTC" firstStartedPulling="2026-02-20 06:58:18.181463781 +0000 UTC m=+1054.952752760" lastFinishedPulling="2026-02-20 06:58:33.69592985 +0000 UTC m=+1070.467218829" observedRunningTime="2026-02-20 06:58:34.799225021 +0000 UTC m=+1071.570513999" watchObservedRunningTime="2026-02-20 06:58:34.804716974 +0000 UTC m=+1071.576005952" Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.834530 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-58b9894566-r5jqg"] Feb 20 06:58:34 crc kubenswrapper[4492]: I0220 06:58:34.865240 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-58b9894566-r5jqg"] Feb 20 06:58:35 crc kubenswrapper[4492]: I0220 06:58:35.567327 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" path="/var/lib/kubelet/pods/e7da01e6-bb65-47dd-8357-72f1fec9d95d/volumes" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.312599 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.403614 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-scripts\") pod \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.403683 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-run-httpd\") pod \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.403825 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5tzh\" (UniqueName: \"kubernetes.io/projected/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-kube-api-access-t5tzh\") pod \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.403883 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-config-data\") pod \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.403917 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-sg-core-conf-yaml\") pod \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.404004 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-log-httpd\") pod \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.404024 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-combined-ca-bundle\") pod \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\" (UID: \"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7\") " Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.414385 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" (UID: "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.416688 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-scripts" (OuterVolumeSpecName: "scripts") pod "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" (UID: "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.425574 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-kube-api-access-t5tzh" (OuterVolumeSpecName: "kube-api-access-t5tzh") pod "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" (UID: "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7"). InnerVolumeSpecName "kube-api-access-t5tzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.428579 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" (UID: "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.457206 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" (UID: "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.491604 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" (UID: "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.507363 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.507393 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.507404 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.507414 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.507422 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5tzh\" (UniqueName: \"kubernetes.io/projected/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-kube-api-access-t5tzh\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.507433 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.509200 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-config-data" (OuterVolumeSpecName: "config-data") pod "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" (UID: "06d315e6-3640-43d7-9e7f-eaa86a0f7eb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.609080 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.861145 4492 generic.go:334] "Generic (PLEG): container finished" podID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerID="2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b" exitCode=0 Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.861197 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerDied","Data":"2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b"} Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.861253 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06d315e6-3640-43d7-9e7f-eaa86a0f7eb7","Type":"ContainerDied","Data":"ed8085cd3de2f8eea4306f8968546d749be441902cafcdf6b3e48cca96dbe040"} Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.861274 4492 scope.go:117] "RemoveContainer" containerID="cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.861497 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.897005 4492 scope.go:117] "RemoveContainer" containerID="7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.897164 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.903832 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.935701 4492 scope.go:117] "RemoveContainer" containerID="6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.979079 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:37 crc kubenswrapper[4492]: E0220 06:58:37.979913 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" containerName="heat-engine" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.979934 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" containerName="heat-engine" Feb 20 06:58:37 crc kubenswrapper[4492]: E0220 06:58:37.979959 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="sg-core" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.979982 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="sg-core" Feb 20 06:58:37 crc kubenswrapper[4492]: E0220 06:58:37.979990 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-notification-agent" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.979996 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-notification-agent" Feb 20 06:58:37 crc kubenswrapper[4492]: E0220 06:58:37.980010 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="proxy-httpd" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.980017 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="proxy-httpd" Feb 20 06:58:37 crc kubenswrapper[4492]: E0220 06:58:37.980041 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-central-agent" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.980066 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-central-agent" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.980322 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="sg-core" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.980340 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="proxy-httpd" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.980347 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-central-agent" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.980385 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7da01e6-bb65-47dd-8357-72f1fec9d95d" containerName="heat-engine" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.980396 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" containerName="ceilometer-notification-agent" Feb 20 06:58:37 crc kubenswrapper[4492]: I0220 06:58:37.986409 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.004205 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.005095 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.008875 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.027667 4492 scope.go:117] "RemoveContainer" containerID="2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.104524 4492 scope.go:117] "RemoveContainer" containerID="cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3" Feb 20 06:58:38 crc kubenswrapper[4492]: E0220 06:58:38.104972 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3\": container with ID starting with cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3 not found: ID does not exist" containerID="cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.105023 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3"} err="failed to get container status \"cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3\": rpc error: code = NotFound desc = could not find container \"cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3\": container with ID starting with cc9771fa70433aeafb3c9e89a078d2b6a661b82bf55c0319a41cb1f43efdf0c3 not found: ID does not exist" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.105043 4492 scope.go:117] "RemoveContainer" containerID="7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c" Feb 20 06:58:38 crc kubenswrapper[4492]: E0220 06:58:38.105764 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c\": container with ID starting with 7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c not found: ID does not exist" containerID="7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.105788 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c"} err="failed to get container status \"7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c\": rpc error: code = NotFound desc = could not find container \"7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c\": container with ID starting with 7a9b5af9d5b1c3ee3fa3863c1a80bb0b6e052a1db0c13932195fa52262c8ab3c not found: ID does not exist" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.105804 4492 scope.go:117] "RemoveContainer" containerID="6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c" Feb 20 06:58:38 crc kubenswrapper[4492]: E0220 06:58:38.106032 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c\": container with ID starting with 6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c not found: ID does not exist" containerID="6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.106049 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c"} err="failed to get container status \"6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c\": rpc error: code = NotFound desc = could not find container \"6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c\": container with ID starting with 6bdb4ea8cef764c7369dfa02c27dbb5f8d389f89bafe40f3fc293abe34d94b3c not found: ID does not exist" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.106081 4492 scope.go:117] "RemoveContainer" containerID="2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b" Feb 20 06:58:38 crc kubenswrapper[4492]: E0220 06:58:38.106285 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b\": container with ID starting with 2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b not found: ID does not exist" containerID="2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.106299 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b"} err="failed to get container status \"2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b\": rpc error: code = NotFound desc = could not find container \"2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b\": container with ID starting with 2d0d337a809e751ba391ee2977601a15b5900955b1c6d02c6ed8c079d204796b not found: ID does not exist" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.130862 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-scripts\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.131155 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.131188 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-config-data\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.131265 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-log-httpd\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.131359 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-run-httpd\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.131468 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.131578 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7v5g\" (UniqueName: \"kubernetes.io/projected/2f77deb8-f79e-49f7-875c-aba45f4b0977-kube-api-access-b7v5g\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.233429 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.233748 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-config-data\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.233786 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-log-httpd\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.233828 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-run-httpd\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.233870 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.233896 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7v5g\" (UniqueName: \"kubernetes.io/projected/2f77deb8-f79e-49f7-875c-aba45f4b0977-kube-api-access-b7v5g\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.233912 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-scripts\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.235032 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-log-httpd\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.235563 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-run-httpd\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.238869 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.240139 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.240732 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-config-data\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.256836 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-scripts\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.260154 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7v5g\" (UniqueName: \"kubernetes.io/projected/2f77deb8-f79e-49f7-875c-aba45f4b0977-kube-api-access-b7v5g\") pod \"ceilometer-0\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.328007 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.808828 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:38 crc kubenswrapper[4492]: W0220 06:58:38.814615 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f77deb8_f79e_49f7_875c_aba45f4b0977.slice/crio-a01c32273248202ea89e1a8231822fb8dc689b7917fbc66d37ca41836e99c1ec WatchSource:0}: Error finding container a01c32273248202ea89e1a8231822fb8dc689b7917fbc66d37ca41836e99c1ec: Status 404 returned error can't find the container with id a01c32273248202ea89e1a8231822fb8dc689b7917fbc66d37ca41836e99c1ec Feb 20 06:58:38 crc kubenswrapper[4492]: I0220 06:58:38.873040 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerStarted","Data":"a01c32273248202ea89e1a8231822fb8dc689b7917fbc66d37ca41836e99c1ec"} Feb 20 06:58:39 crc kubenswrapper[4492]: I0220 06:58:39.577597 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06d315e6-3640-43d7-9e7f-eaa86a0f7eb7" path="/var/lib/kubelet/pods/06d315e6-3640-43d7-9e7f-eaa86a0f7eb7/volumes" Feb 20 06:58:39 crc kubenswrapper[4492]: I0220 06:58:39.885808 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerStarted","Data":"b8bcadccc077a072ddf622973e188bafb6c91e58b1586239ec90786d7b7cad7a"} Feb 20 06:58:40 crc kubenswrapper[4492]: I0220 06:58:40.897294 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerStarted","Data":"ec29179caa35b44ea7c3fd4cefbb1301d7d2414919c86493041d341ae008746c"} Feb 20 06:58:41 crc kubenswrapper[4492]: I0220 06:58:41.908119 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerStarted","Data":"9872e866b203c8b65f2f4650d664f7d3d7e29ec00a9f07461b81a2f5e3a19adb"} Feb 20 06:58:42 crc kubenswrapper[4492]: I0220 06:58:42.924793 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerStarted","Data":"50b83c0ff9608ee76e0d6bfbb33906d4cb0567180049e7d00ffc2c3d6acc95d6"} Feb 20 06:58:42 crc kubenswrapper[4492]: I0220 06:58:42.925330 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 06:58:42 crc kubenswrapper[4492]: I0220 06:58:42.951410 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.53702876 podStartE2EDuration="5.951389813s" podCreationTimestamp="2026-02-20 06:58:37 +0000 UTC" firstStartedPulling="2026-02-20 06:58:38.817208599 +0000 UTC m=+1075.588497577" lastFinishedPulling="2026-02-20 06:58:42.231569652 +0000 UTC m=+1079.002858630" observedRunningTime="2026-02-20 06:58:42.942940124 +0000 UTC m=+1079.714229103" watchObservedRunningTime="2026-02-20 06:58:42.951389813 +0000 UTC m=+1079.722678791" Feb 20 06:58:43 crc kubenswrapper[4492]: I0220 06:58:43.940823 4492 generic.go:334] "Generic (PLEG): container finished" podID="d97b093e-5584-4662-a976-215ada9ef003" containerID="7f0f8e79841d1114a5bd817e65c94bd327be8a4c19392bc146b97a3c66ec3466" exitCode=0 Feb 20 06:58:43 crc kubenswrapper[4492]: I0220 06:58:43.943571 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" event={"ID":"d97b093e-5584-4662-a976-215ada9ef003","Type":"ContainerDied","Data":"7f0f8e79841d1114a5bd817e65c94bd327be8a4c19392bc146b97a3c66ec3466"} Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.256380 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.392300 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-config-data\") pod \"d97b093e-5584-4662-a976-215ada9ef003\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.392372 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-scripts\") pod \"d97b093e-5584-4662-a976-215ada9ef003\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.392427 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29wlr\" (UniqueName: \"kubernetes.io/projected/d97b093e-5584-4662-a976-215ada9ef003-kube-api-access-29wlr\") pod \"d97b093e-5584-4662-a976-215ada9ef003\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.392551 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-combined-ca-bundle\") pod \"d97b093e-5584-4662-a976-215ada9ef003\" (UID: \"d97b093e-5584-4662-a976-215ada9ef003\") " Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.400122 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d97b093e-5584-4662-a976-215ada9ef003-kube-api-access-29wlr" (OuterVolumeSpecName: "kube-api-access-29wlr") pod "d97b093e-5584-4662-a976-215ada9ef003" (UID: "d97b093e-5584-4662-a976-215ada9ef003"). InnerVolumeSpecName "kube-api-access-29wlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.410856 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-scripts" (OuterVolumeSpecName: "scripts") pod "d97b093e-5584-4662-a976-215ada9ef003" (UID: "d97b093e-5584-4662-a976-215ada9ef003"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.431610 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-config-data" (OuterVolumeSpecName: "config-data") pod "d97b093e-5584-4662-a976-215ada9ef003" (UID: "d97b093e-5584-4662-a976-215ada9ef003"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.453621 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d97b093e-5584-4662-a976-215ada9ef003" (UID: "d97b093e-5584-4662-a976-215ada9ef003"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.494818 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.494937 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.495004 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d97b093e-5584-4662-a976-215ada9ef003-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.495058 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29wlr\" (UniqueName: \"kubernetes.io/projected/d97b093e-5584-4662-a976-215ada9ef003-kube-api-access-29wlr\") on node \"crc\" DevicePath \"\"" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.958769 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" event={"ID":"d97b093e-5584-4662-a976-215ada9ef003","Type":"ContainerDied","Data":"320bb74a3f71d9d1a7833b863ea38235da98818e7439f04b1a6ea4fe66e62284"} Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.958836 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="320bb74a3f71d9d1a7833b863ea38235da98818e7439f04b1a6ea4fe66e62284" Feb 20 06:58:45 crc kubenswrapper[4492]: I0220 06:58:45.958832 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-bsh7r" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.114497 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 06:58:46 crc kubenswrapper[4492]: E0220 06:58:46.118045 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d97b093e-5584-4662-a976-215ada9ef003" containerName="nova-cell0-conductor-db-sync" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.118063 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d97b093e-5584-4662-a976-215ada9ef003" containerName="nova-cell0-conductor-db-sync" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.118263 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d97b093e-5584-4662-a976-215ada9ef003" containerName="nova-cell0-conductor-db-sync" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.118852 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.132186 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.136489 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-xpp5p" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.136694 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.209254 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb65e8c-429a-4279-a277-3730fa1fcd42-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.209345 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb65e8c-429a-4279-a277-3730fa1fcd42-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.209429 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpggp\" (UniqueName: \"kubernetes.io/projected/8fb65e8c-429a-4279-a277-3730fa1fcd42-kube-api-access-kpggp\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.311826 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb65e8c-429a-4279-a277-3730fa1fcd42-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.311939 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpggp\" (UniqueName: \"kubernetes.io/projected/8fb65e8c-429a-4279-a277-3730fa1fcd42-kube-api-access-kpggp\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.312032 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb65e8c-429a-4279-a277-3730fa1fcd42-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.320173 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fb65e8c-429a-4279-a277-3730fa1fcd42-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.320845 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fb65e8c-429a-4279-a277-3730fa1fcd42-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.334908 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpggp\" (UniqueName: \"kubernetes.io/projected/8fb65e8c-429a-4279-a277-3730fa1fcd42-kube-api-access-kpggp\") pod \"nova-cell0-conductor-0\" (UID: \"8fb65e8c-429a-4279-a277-3730fa1fcd42\") " pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.446043 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.893370 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 20 06:58:46 crc kubenswrapper[4492]: W0220 06:58:46.905914 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fb65e8c_429a_4279_a277_3730fa1fcd42.slice/crio-0aebff0d2f97a59e67e989b9981a0f07f668feab64ad643235840085684c5af8 WatchSource:0}: Error finding container 0aebff0d2f97a59e67e989b9981a0f07f668feab64ad643235840085684c5af8: Status 404 returned error can't find the container with id 0aebff0d2f97a59e67e989b9981a0f07f668feab64ad643235840085684c5af8 Feb 20 06:58:46 crc kubenswrapper[4492]: I0220 06:58:46.970521 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8fb65e8c-429a-4279-a277-3730fa1fcd42","Type":"ContainerStarted","Data":"0aebff0d2f97a59e67e989b9981a0f07f668feab64ad643235840085684c5af8"} Feb 20 06:58:47 crc kubenswrapper[4492]: I0220 06:58:47.989227 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8fb65e8c-429a-4279-a277-3730fa1fcd42","Type":"ContainerStarted","Data":"0a8925905e3e19ff30829f166803fc822624202df6a13008c45460dc11df4d8c"} Feb 20 06:58:47 crc kubenswrapper[4492]: I0220 06:58:47.992217 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:48 crc kubenswrapper[4492]: I0220 06:58:48.017872 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.017857201 podStartE2EDuration="2.017857201s" podCreationTimestamp="2026-02-20 06:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:48.010285447 +0000 UTC m=+1084.781574426" watchObservedRunningTime="2026-02-20 06:58:48.017857201 +0000 UTC m=+1084.789146179" Feb 20 06:58:55 crc kubenswrapper[4492]: I0220 06:58:55.768611 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:58:55 crc kubenswrapper[4492]: I0220 06:58:55.770240 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="sg-core" containerID="cri-o://9872e866b203c8b65f2f4650d664f7d3d7e29ec00a9f07461b81a2f5e3a19adb" gracePeriod=30 Feb 20 06:58:55 crc kubenswrapper[4492]: I0220 06:58:55.770363 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-central-agent" containerID="cri-o://b8bcadccc077a072ddf622973e188bafb6c91e58b1586239ec90786d7b7cad7a" gracePeriod=30 Feb 20 06:58:55 crc kubenswrapper[4492]: I0220 06:58:55.770578 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="proxy-httpd" containerID="cri-o://50b83c0ff9608ee76e0d6bfbb33906d4cb0567180049e7d00ffc2c3d6acc95d6" gracePeriod=30 Feb 20 06:58:55 crc kubenswrapper[4492]: I0220 06:58:55.771066 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-notification-agent" containerID="cri-o://ec29179caa35b44ea7c3fd4cefbb1301d7d2414919c86493041d341ae008746c" gracePeriod=30 Feb 20 06:58:55 crc kubenswrapper[4492]: I0220 06:58:55.788708 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.198:3000/\": EOF" Feb 20 06:58:56 crc kubenswrapper[4492]: I0220 06:58:56.081437 4492 generic.go:334] "Generic (PLEG): container finished" podID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerID="50b83c0ff9608ee76e0d6bfbb33906d4cb0567180049e7d00ffc2c3d6acc95d6" exitCode=0 Feb 20 06:58:56 crc kubenswrapper[4492]: I0220 06:58:56.081866 4492 generic.go:334] "Generic (PLEG): container finished" podID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerID="9872e866b203c8b65f2f4650d664f7d3d7e29ec00a9f07461b81a2f5e3a19adb" exitCode=2 Feb 20 06:58:56 crc kubenswrapper[4492]: I0220 06:58:56.081532 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerDied","Data":"50b83c0ff9608ee76e0d6bfbb33906d4cb0567180049e7d00ffc2c3d6acc95d6"} Feb 20 06:58:56 crc kubenswrapper[4492]: I0220 06:58:56.081925 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerDied","Data":"9872e866b203c8b65f2f4650d664f7d3d7e29ec00a9f07461b81a2f5e3a19adb"} Feb 20 06:58:56 crc kubenswrapper[4492]: I0220 06:58:56.469964 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.038434 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-bt5pz"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.041922 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.047873 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.048054 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.058158 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bt5pz"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.092451 4492 generic.go:334] "Generic (PLEG): container finished" podID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerID="b8bcadccc077a072ddf622973e188bafb6c91e58b1586239ec90786d7b7cad7a" exitCode=0 Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.092517 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerDied","Data":"b8bcadccc077a072ddf622973e188bafb6c91e58b1586239ec90786d7b7cad7a"} Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.144737 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6lll\" (UniqueName: \"kubernetes.io/projected/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-kube-api-access-n6lll\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.144884 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-scripts\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.145181 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.145322 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-config-data\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.227167 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.228624 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.232604 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.247900 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6lll\" (UniqueName: \"kubernetes.io/projected/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-kube-api-access-n6lll\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.248043 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-scripts\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.248179 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.248216 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-config-data\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.260009 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-scripts\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.270020 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-config-data\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.272730 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.296215 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.298091 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.310746 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6lll\" (UniqueName: \"kubernetes.io/projected/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-kube-api-access-n6lll\") pod \"nova-cell0-cell-mapping-bt5pz\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.323934 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.330886 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.354820 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wptlh\" (UniqueName: \"kubernetes.io/projected/31b6bf8f-6008-42dc-9ebf-a28877958e40-kube-api-access-wptlh\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.354896 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b6bf8f-6008-42dc-9ebf-a28877958e40-logs\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.354941 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-config-data\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.354998 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.355018 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-logs\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.355046 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-config-data\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.355044 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.355063 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdqdd\" (UniqueName: \"kubernetes.io/projected/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-kube-api-access-fdqdd\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.355091 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.356445 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.357120 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.364714 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.372812 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.398216 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.414217 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.429059 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457543 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-config-data\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457584 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdqdd\" (UniqueName: \"kubernetes.io/projected/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-kube-api-access-fdqdd\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457635 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457657 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457779 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wptlh\" (UniqueName: \"kubernetes.io/projected/31b6bf8f-6008-42dc-9ebf-a28877958e40-kube-api-access-wptlh\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457806 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457843 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwfc4\" (UniqueName: \"kubernetes.io/projected/fdf044a0-70f6-4682-a663-57b29bc74a17-kube-api-access-wwfc4\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457895 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b6bf8f-6008-42dc-9ebf-a28877958e40-logs\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.457970 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-config-data\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.458002 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.458017 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.458046 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cpq7\" (UniqueName: \"kubernetes.io/projected/7e4a8624-84f4-44bb-b7be-7f73c168e78b-kube-api-access-7cpq7\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.458113 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.458132 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-logs\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.458512 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-logs\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.466016 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.466406 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b6bf8f-6008-42dc-9ebf-a28877958e40-logs\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.477197 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-config-data\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.483980 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-config-data\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.485493 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.486790 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.493225 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wptlh\" (UniqueName: \"kubernetes.io/projected/31b6bf8f-6008-42dc-9ebf-a28877958e40-kube-api-access-wptlh\") pod \"nova-api-0\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.494176 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdqdd\" (UniqueName: \"kubernetes.io/projected/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-kube-api-access-fdqdd\") pod \"nova-metadata-0\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.535393 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.551596 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.559460 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.561156 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.561190 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.561215 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cpq7\" (UniqueName: \"kubernetes.io/projected/7e4a8624-84f4-44bb-b7be-7f73c168e78b-kube-api-access-7cpq7\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.561284 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.561354 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.561381 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwfc4\" (UniqueName: \"kubernetes.io/projected/fdf044a0-70f6-4682-a663-57b29bc74a17-kube-api-access-wwfc4\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.568587 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.571144 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.573445 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.574103 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.595897 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwfc4\" (UniqueName: \"kubernetes.io/projected/fdf044a0-70f6-4682-a663-57b29bc74a17-kube-api-access-wwfc4\") pod \"nova-cell1-novncproxy-0\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.602611 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cpq7\" (UniqueName: \"kubernetes.io/projected/7e4a8624-84f4-44bb-b7be-7f73c168e78b-kube-api-access-7cpq7\") pod \"nova-scheduler-0\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.689238 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.701188 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.710530 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b946fdf6f-dz5bm"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.715439 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b946fdf6f-dz5bm"] Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.715567 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.780646 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-nb\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.780760 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d555\" (UniqueName: \"kubernetes.io/projected/26c40a07-b59c-4796-a5cb-caf6d8482de2-kube-api-access-5d555\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.780874 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-svc\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.780904 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-swift-storage-0\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.781062 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-config\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.781129 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-sb\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.882799 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-nb\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.882888 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d555\" (UniqueName: \"kubernetes.io/projected/26c40a07-b59c-4796-a5cb-caf6d8482de2-kube-api-access-5d555\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.882936 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-svc\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.882956 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-swift-storage-0\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.883035 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-config\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.883082 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-sb\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.883963 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-sb\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.884486 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-nb\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.885304 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-svc\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.885447 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-swift-storage-0\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.886432 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-config\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:57 crc kubenswrapper[4492]: I0220 06:58:57.906098 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d555\" (UniqueName: \"kubernetes.io/projected/26c40a07-b59c-4796-a5cb-caf6d8482de2-kube-api-access-5d555\") pod \"dnsmasq-dns-b946fdf6f-dz5bm\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.054282 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.095398 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-bt5pz"] Feb 20 06:58:58 crc kubenswrapper[4492]: W0220 06:58:58.102568 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0a2bbcf_c2d7_47b7_873d_4b93d85c963a.slice/crio-bfa79cfbdd98067149873aa0ca00100152eb637b7d7d602a1f377bc12dc8d39c WatchSource:0}: Error finding container bfa79cfbdd98067149873aa0ca00100152eb637b7d7d602a1f377bc12dc8d39c: Status 404 returned error can't find the container with id bfa79cfbdd98067149873aa0ca00100152eb637b7d7d602a1f377bc12dc8d39c Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.205946 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.313909 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.399716 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.451635 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.494291 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b946fdf6f-dz5bm"] Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.585314 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rj7nh"] Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.586652 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.590601 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.590842 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.610420 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rj7nh"] Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.700608 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-config-data\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.700660 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-scripts\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.700745 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.700896 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggtlh\" (UniqueName: \"kubernetes.io/projected/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-kube-api-access-ggtlh\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.803064 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-config-data\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.803115 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-scripts\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.803153 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.803977 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggtlh\" (UniqueName: \"kubernetes.io/projected/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-kube-api-access-ggtlh\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.809396 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-scripts\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.809528 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-config-data\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.809928 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.822572 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggtlh\" (UniqueName: \"kubernetes.io/projected/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-kube-api-access-ggtlh\") pod \"nova-cell1-conductor-db-sync-rj7nh\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:58 crc kubenswrapper[4492]: I0220 06:58:58.905362 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.121564 4492 generic.go:334] "Generic (PLEG): container finished" podID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerID="37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e" exitCode=0 Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.121866 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" event={"ID":"26c40a07-b59c-4796-a5cb-caf6d8482de2","Type":"ContainerDied","Data":"37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.121893 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" event={"ID":"26c40a07-b59c-4796-a5cb-caf6d8482de2","Type":"ContainerStarted","Data":"6c7d1e8174755ecfcf503de1b6cb3b17bf95b81e05f503f868a7e1dbe237445f"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.127571 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31b6bf8f-6008-42dc-9ebf-a28877958e40","Type":"ContainerStarted","Data":"0970aef915bd8f1d85ccaa0912930ae3e6312b2a6a850dab94e2eceea1de2a48"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.142004 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fdf044a0-70f6-4682-a663-57b29bc74a17","Type":"ContainerStarted","Data":"467634700eb2f37c2fdc9f41873d7d0c3a28e9eee6deac89065d1460d27c1bf3"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.172753 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e4a8624-84f4-44bb-b7be-7f73c168e78b","Type":"ContainerStarted","Data":"3ad64b8880da47a00a27934a426c81a5e29c475b1163452fa8ef79b89249ad75"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.197065 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bt5pz" event={"ID":"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a","Type":"ContainerStarted","Data":"87e591d8080db489d7bfc1c3018db2736c372426f84b947b74109bffb920aba8"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.197215 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bt5pz" event={"ID":"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a","Type":"ContainerStarted","Data":"bfa79cfbdd98067149873aa0ca00100152eb637b7d7d602a1f377bc12dc8d39c"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.200792 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa","Type":"ContainerStarted","Data":"c3cf1de1c39ebac3508e337f377bc7adc50cbb4c734a3d857aea9161768a9d56"} Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.224294 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-bt5pz" podStartSLOduration=2.224276786 podStartE2EDuration="2.224276786s" podCreationTimestamp="2026-02-20 06:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:58:59.212870013 +0000 UTC m=+1095.984158992" watchObservedRunningTime="2026-02-20 06:58:59.224276786 +0000 UTC m=+1095.995565763" Feb 20 06:58:59 crc kubenswrapper[4492]: I0220 06:58:59.364990 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rj7nh"] Feb 20 06:58:59 crc kubenswrapper[4492]: W0220 06:58:59.414231 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4c9cd39_beac_4a57_ba33_ae3f41187e8e.slice/crio-580434c0ce39142ac30f20526ebcd65523cbfbdec4c8598f86ed25eeaeabed3b WatchSource:0}: Error finding container 580434c0ce39142ac30f20526ebcd65523cbfbdec4c8598f86ed25eeaeabed3b: Status 404 returned error can't find the container with id 580434c0ce39142ac30f20526ebcd65523cbfbdec4c8598f86ed25eeaeabed3b Feb 20 06:59:00 crc kubenswrapper[4492]: I0220 06:59:00.225637 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" event={"ID":"26c40a07-b59c-4796-a5cb-caf6d8482de2","Type":"ContainerStarted","Data":"4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550"} Feb 20 06:59:00 crc kubenswrapper[4492]: I0220 06:59:00.227622 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:59:00 crc kubenswrapper[4492]: I0220 06:59:00.237652 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" event={"ID":"a4c9cd39-beac-4a57-ba33-ae3f41187e8e","Type":"ContainerStarted","Data":"aa0246f17b0e5dd8261abc5bd0187312e71fe50e6481202bf3406d51922e0faa"} Feb 20 06:59:00 crc kubenswrapper[4492]: I0220 06:59:00.237862 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" event={"ID":"a4c9cd39-beac-4a57-ba33-ae3f41187e8e","Type":"ContainerStarted","Data":"580434c0ce39142ac30f20526ebcd65523cbfbdec4c8598f86ed25eeaeabed3b"} Feb 20 06:59:00 crc kubenswrapper[4492]: I0220 06:59:00.266419 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" podStartSLOduration=3.266400887 podStartE2EDuration="3.266400887s" podCreationTimestamp="2026-02-20 06:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:00.248815008 +0000 UTC m=+1097.020103986" watchObservedRunningTime="2026-02-20 06:59:00.266400887 +0000 UTC m=+1097.037689865" Feb 20 06:59:00 crc kubenswrapper[4492]: I0220 06:59:00.273985 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" podStartSLOduration=2.273971658 podStartE2EDuration="2.273971658s" podCreationTimestamp="2026-02-20 06:58:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:00.264634057 +0000 UTC m=+1097.035923035" watchObservedRunningTime="2026-02-20 06:59:00.273971658 +0000 UTC m=+1097.045260637" Feb 20 06:59:01 crc kubenswrapper[4492]: I0220 06:59:01.247094 4492 generic.go:334] "Generic (PLEG): container finished" podID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerID="ec29179caa35b44ea7c3fd4cefbb1301d7d2414919c86493041d341ae008746c" exitCode=0 Feb 20 06:59:01 crc kubenswrapper[4492]: I0220 06:59:01.247172 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerDied","Data":"ec29179caa35b44ea7c3fd4cefbb1301d7d2414919c86493041d341ae008746c"} Feb 20 06:59:01 crc kubenswrapper[4492]: I0220 06:59:01.616144 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:59:01 crc kubenswrapper[4492]: I0220 06:59:01.644922 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.293108 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2f77deb8-f79e-49f7-875c-aba45f4b0977","Type":"ContainerDied","Data":"a01c32273248202ea89e1a8231822fb8dc689b7917fbc66d37ca41836e99c1ec"} Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.293355 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a01c32273248202ea89e1a8231822fb8dc689b7917fbc66d37ca41836e99c1ec" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.293542 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.486404 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-run-httpd\") pod \"2f77deb8-f79e-49f7-875c-aba45f4b0977\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.487504 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-combined-ca-bundle\") pod \"2f77deb8-f79e-49f7-875c-aba45f4b0977\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.487067 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2f77deb8-f79e-49f7-875c-aba45f4b0977" (UID: "2f77deb8-f79e-49f7-875c-aba45f4b0977"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.487746 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-scripts\") pod \"2f77deb8-f79e-49f7-875c-aba45f4b0977\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.487781 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-config-data\") pod \"2f77deb8-f79e-49f7-875c-aba45f4b0977\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.488607 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7v5g\" (UniqueName: \"kubernetes.io/projected/2f77deb8-f79e-49f7-875c-aba45f4b0977-kube-api-access-b7v5g\") pod \"2f77deb8-f79e-49f7-875c-aba45f4b0977\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.490129 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-sg-core-conf-yaml\") pod \"2f77deb8-f79e-49f7-875c-aba45f4b0977\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.490408 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-log-httpd\") pod \"2f77deb8-f79e-49f7-875c-aba45f4b0977\" (UID: \"2f77deb8-f79e-49f7-875c-aba45f4b0977\") " Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.492865 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.494217 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2f77deb8-f79e-49f7-875c-aba45f4b0977" (UID: "2f77deb8-f79e-49f7-875c-aba45f4b0977"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.497918 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f77deb8-f79e-49f7-875c-aba45f4b0977-kube-api-access-b7v5g" (OuterVolumeSpecName: "kube-api-access-b7v5g") pod "2f77deb8-f79e-49f7-875c-aba45f4b0977" (UID: "2f77deb8-f79e-49f7-875c-aba45f4b0977"). InnerVolumeSpecName "kube-api-access-b7v5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.504773 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-scripts" (OuterVolumeSpecName: "scripts") pod "2f77deb8-f79e-49f7-875c-aba45f4b0977" (UID: "2f77deb8-f79e-49f7-875c-aba45f4b0977"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.536394 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2f77deb8-f79e-49f7-875c-aba45f4b0977" (UID: "2f77deb8-f79e-49f7-875c-aba45f4b0977"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.595155 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.595188 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7v5g\" (UniqueName: \"kubernetes.io/projected/2f77deb8-f79e-49f7-875c-aba45f4b0977-kube-api-access-b7v5g\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.595199 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.595211 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2f77deb8-f79e-49f7-875c-aba45f4b0977-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.679584 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f77deb8-f79e-49f7-875c-aba45f4b0977" (UID: "2f77deb8-f79e-49f7-875c-aba45f4b0977"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.698439 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.704607 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-config-data" (OuterVolumeSpecName: "config-data") pod "2f77deb8-f79e-49f7-875c-aba45f4b0977" (UID: "2f77deb8-f79e-49f7-875c-aba45f4b0977"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:03 crc kubenswrapper[4492]: I0220 06:59:03.800266 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f77deb8-f79e-49f7-875c-aba45f4b0977-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.309345 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fdf044a0-70f6-4682-a663-57b29bc74a17","Type":"ContainerStarted","Data":"51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a"} Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.309538 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fdf044a0-70f6-4682-a663-57b29bc74a17" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a" gracePeriod=30 Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.315290 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e4a8624-84f4-44bb-b7be-7f73c168e78b","Type":"ContainerStarted","Data":"7a5565002e776eb16a686355c427604138461502990fdfb251804652f7a0583e"} Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.333755 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa","Type":"ContainerStarted","Data":"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d"} Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.333808 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa","Type":"ContainerStarted","Data":"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df"} Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.333931 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-log" containerID="cri-o://06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df" gracePeriod=30 Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.334067 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-metadata" containerID="cri-o://3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d" gracePeriod=30 Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.335009 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.5378849580000002 podStartE2EDuration="7.334992504s" podCreationTimestamp="2026-02-20 06:58:57 +0000 UTC" firstStartedPulling="2026-02-20 06:58:58.403068396 +0000 UTC m=+1095.174357374" lastFinishedPulling="2026-02-20 06:59:03.200175943 +0000 UTC m=+1099.971464920" observedRunningTime="2026-02-20 06:59:04.324864591 +0000 UTC m=+1101.096153570" watchObservedRunningTime="2026-02-20 06:59:04.334992504 +0000 UTC m=+1101.106281481" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.345563 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.345765 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31b6bf8f-6008-42dc-9ebf-a28877958e40","Type":"ContainerStarted","Data":"340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f"} Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.345814 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31b6bf8f-6008-42dc-9ebf-a28877958e40","Type":"ContainerStarted","Data":"bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622"} Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.374234 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.595493109 podStartE2EDuration="7.374206001s" podCreationTimestamp="2026-02-20 06:58:57 +0000 UTC" firstStartedPulling="2026-02-20 06:58:58.398580946 +0000 UTC m=+1095.169869925" lastFinishedPulling="2026-02-20 06:59:03.177293839 +0000 UTC m=+1099.948582817" observedRunningTime="2026-02-20 06:59:04.351207589 +0000 UTC m=+1101.122496566" watchObservedRunningTime="2026-02-20 06:59:04.374206001 +0000 UTC m=+1101.145494979" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.382255 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.422774729 podStartE2EDuration="7.382228985s" podCreationTimestamp="2026-02-20 06:58:57 +0000 UTC" firstStartedPulling="2026-02-20 06:58:58.220138737 +0000 UTC m=+1094.991427716" lastFinishedPulling="2026-02-20 06:59:03.179592994 +0000 UTC m=+1099.950881972" observedRunningTime="2026-02-20 06:59:04.372915108 +0000 UTC m=+1101.144204086" watchObservedRunningTime="2026-02-20 06:59:04.382228985 +0000 UTC m=+1101.153517963" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.397395 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.552272515 podStartE2EDuration="7.397387468s" podCreationTimestamp="2026-02-20 06:58:57 +0000 UTC" firstStartedPulling="2026-02-20 06:58:58.332175941 +0000 UTC m=+1095.103464919" lastFinishedPulling="2026-02-20 06:59:03.177290894 +0000 UTC m=+1099.948579872" observedRunningTime="2026-02-20 06:59:04.385912809 +0000 UTC m=+1101.157201787" watchObservedRunningTime="2026-02-20 06:59:04.397387468 +0000 UTC m=+1101.168676437" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.428935 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.435531 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.447526 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:04 crc kubenswrapper[4492]: E0220 06:59:04.448052 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-notification-agent" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448068 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-notification-agent" Feb 20 06:59:04 crc kubenswrapper[4492]: E0220 06:59:04.448096 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="sg-core" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448103 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="sg-core" Feb 20 06:59:04 crc kubenswrapper[4492]: E0220 06:59:04.448120 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="proxy-httpd" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448128 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="proxy-httpd" Feb 20 06:59:04 crc kubenswrapper[4492]: E0220 06:59:04.448139 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-central-agent" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448145 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-central-agent" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448343 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-central-agent" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448356 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="ceilometer-notification-agent" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448378 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="proxy-httpd" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.448392 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" containerName="sg-core" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.450124 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.456194 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.462981 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.463128 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.618024 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-run-httpd\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.618303 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.618562 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-scripts\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.618776 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4c6x\" (UniqueName: \"kubernetes.io/projected/a995bed7-d939-4344-bac8-b506a197e28f-kube-api-access-k4c6x\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.618817 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-log-httpd\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.618864 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.618912 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-config-data\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.721240 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-scripts\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.722143 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4c6x\" (UniqueName: \"kubernetes.io/projected/a995bed7-d939-4344-bac8-b506a197e28f-kube-api-access-k4c6x\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.722208 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-log-httpd\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.722279 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.722342 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-config-data\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.722493 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-run-httpd\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.722532 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.723065 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-log-httpd\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.723296 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-run-httpd\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.742871 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4c6x\" (UniqueName: \"kubernetes.io/projected/a995bed7-d939-4344-bac8-b506a197e28f-kube-api-access-k4c6x\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.751301 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-config-data\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.753317 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.762721 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.778001 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-scripts\") pod \"ceilometer-0\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.806376 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.806380 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.939573 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-combined-ca-bundle\") pod \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.939802 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-logs\") pod \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.939847 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdqdd\" (UniqueName: \"kubernetes.io/projected/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-kube-api-access-fdqdd\") pod \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.939898 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-config-data\") pod \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\" (UID: \"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa\") " Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.940283 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-logs" (OuterVolumeSpecName: "logs") pod "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" (UID: "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.940624 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.944660 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-kube-api-access-fdqdd" (OuterVolumeSpecName: "kube-api-access-fdqdd") pod "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" (UID: "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa"). InnerVolumeSpecName "kube-api-access-fdqdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.967338 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-config-data" (OuterVolumeSpecName: "config-data") pod "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" (UID: "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:04 crc kubenswrapper[4492]: I0220 06:59:04.981962 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" (UID: "eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.051575 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.051699 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdqdd\" (UniqueName: \"kubernetes.io/projected/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-kube-api-access-fdqdd\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.051765 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.287839 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.361266 4492 generic.go:334] "Generic (PLEG): container finished" podID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerID="3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d" exitCode=0 Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.361310 4492 generic.go:334] "Generic (PLEG): container finished" podID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerID="06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df" exitCode=143 Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.361378 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa","Type":"ContainerDied","Data":"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d"} Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.361412 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa","Type":"ContainerDied","Data":"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df"} Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.361424 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa","Type":"ContainerDied","Data":"c3cf1de1c39ebac3508e337f377bc7adc50cbb4c734a3d857aea9161768a9d56"} Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.361443 4492 scope.go:117] "RemoveContainer" containerID="3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.361660 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.364152 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerStarted","Data":"03dcb71bec1296c043fabf557ef7d12deb827d612d472e894313d06980067df5"} Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.397288 4492 scope.go:117] "RemoveContainer" containerID="06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.420341 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.434524 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.452705 4492 scope.go:117] "RemoveContainer" containerID="3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d" Feb 20 06:59:05 crc kubenswrapper[4492]: E0220 06:59:05.455916 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d\": container with ID starting with 3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d not found: ID does not exist" containerID="3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.455975 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d"} err="failed to get container status \"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d\": rpc error: code = NotFound desc = could not find container \"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d\": container with ID starting with 3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d not found: ID does not exist" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.456009 4492 scope.go:117] "RemoveContainer" containerID="06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df" Feb 20 06:59:05 crc kubenswrapper[4492]: E0220 06:59:05.461605 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df\": container with ID starting with 06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df not found: ID does not exist" containerID="06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.461646 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df"} err="failed to get container status \"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df\": rpc error: code = NotFound desc = could not find container \"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df\": container with ID starting with 06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df not found: ID does not exist" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.461674 4492 scope.go:117] "RemoveContainer" containerID="3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.463043 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d"} err="failed to get container status \"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d\": rpc error: code = NotFound desc = could not find container \"3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d\": container with ID starting with 3a2876d8c0c274d825fc0da3525d338e0fb75b669c3bc89703d84e7f5cb13b7d not found: ID does not exist" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.463070 4492 scope.go:117] "RemoveContainer" containerID="06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.466567 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df"} err="failed to get container status \"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df\": rpc error: code = NotFound desc = could not find container \"06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df\": container with ID starting with 06062b4da4a04d3ee6c762d262fd7b6c74320bf3f8ecb6216e23babf481fe7df not found: ID does not exist" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.467193 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:05 crc kubenswrapper[4492]: E0220 06:59:05.467744 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-metadata" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.467765 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-metadata" Feb 20 06:59:05 crc kubenswrapper[4492]: E0220 06:59:05.467802 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-log" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.467809 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-log" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.468019 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-log" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.468045 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" containerName="nova-metadata-metadata" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.469265 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.475218 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.475383 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.476296 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.568557 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f77deb8-f79e-49f7-875c-aba45f4b0977" path="/var/lib/kubelet/pods/2f77deb8-f79e-49f7-875c-aba45f4b0977/volumes" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.569403 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa" path="/var/lib/kubelet/pods/eb172c2b-d5c7-4bab-91b8-f5c26bd2e4fa/volumes" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.579545 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f53582-925f-4014-a132-5ba0d831f8c9-logs\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.579592 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.579628 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.579688 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-config-data\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.579721 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ljjm\" (UniqueName: \"kubernetes.io/projected/e9f53582-925f-4014-a132-5ba0d831f8c9-kube-api-access-5ljjm\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.681966 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ljjm\" (UniqueName: \"kubernetes.io/projected/e9f53582-925f-4014-a132-5ba0d831f8c9-kube-api-access-5ljjm\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.682218 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f53582-925f-4014-a132-5ba0d831f8c9-logs\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.682304 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.682399 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.682538 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-config-data\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.683110 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f53582-925f-4014-a132-5ba0d831f8c9-logs\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.686997 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.687392 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.690986 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-config-data\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.699746 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ljjm\" (UniqueName: \"kubernetes.io/projected/e9f53582-925f-4014-a132-5ba0d831f8c9-kube-api-access-5ljjm\") pod \"nova-metadata-0\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " pod="openstack/nova-metadata-0" Feb 20 06:59:05 crc kubenswrapper[4492]: I0220 06:59:05.784571 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:06 crc kubenswrapper[4492]: I0220 06:59:06.261381 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:06 crc kubenswrapper[4492]: I0220 06:59:06.372866 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerStarted","Data":"867d954a3a55c27b9f5776f996866986d5d8bbe2fb428dc7d5317e2681b797e2"} Feb 20 06:59:06 crc kubenswrapper[4492]: I0220 06:59:06.375030 4492 generic.go:334] "Generic (PLEG): container finished" podID="a4c9cd39-beac-4a57-ba33-ae3f41187e8e" containerID="aa0246f17b0e5dd8261abc5bd0187312e71fe50e6481202bf3406d51922e0faa" exitCode=0 Feb 20 06:59:06 crc kubenswrapper[4492]: I0220 06:59:06.375100 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" event={"ID":"a4c9cd39-beac-4a57-ba33-ae3f41187e8e","Type":"ContainerDied","Data":"aa0246f17b0e5dd8261abc5bd0187312e71fe50e6481202bf3406d51922e0faa"} Feb 20 06:59:06 crc kubenswrapper[4492]: I0220 06:59:06.377970 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9f53582-925f-4014-a132-5ba0d831f8c9","Type":"ContainerStarted","Data":"207923d592afa28b960748e40202c6761ae09f2e67a7fe608360a9c0a7ac80b4"} Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.391613 4492 generic.go:334] "Generic (PLEG): container finished" podID="f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" containerID="87e591d8080db489d7bfc1c3018db2736c372426f84b947b74109bffb920aba8" exitCode=0 Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.391708 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bt5pz" event={"ID":"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a","Type":"ContainerDied","Data":"87e591d8080db489d7bfc1c3018db2736c372426f84b947b74109bffb920aba8"} Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.395846 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9f53582-925f-4014-a132-5ba0d831f8c9","Type":"ContainerStarted","Data":"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c"} Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.395904 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9f53582-925f-4014-a132-5ba0d831f8c9","Type":"ContainerStarted","Data":"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c"} Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.409970 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerStarted","Data":"155c202ecab9200b503f2a5d96bc709966f2357f1dca077be158f8c3a783fcae"} Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.454592 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.454572585 podStartE2EDuration="2.454572585s" podCreationTimestamp="2026-02-20 06:59:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:07.453786985 +0000 UTC m=+1104.225075953" watchObservedRunningTime="2026-02-20 06:59:07.454572585 +0000 UTC m=+1104.225861564" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.576443 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.576505 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.690408 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.690451 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.701678 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.755666 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.899842 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.931392 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-scripts\") pod \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.931464 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-combined-ca-bundle\") pod \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.931537 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-config-data\") pod \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.931578 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggtlh\" (UniqueName: \"kubernetes.io/projected/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-kube-api-access-ggtlh\") pod \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\" (UID: \"a4c9cd39-beac-4a57-ba33-ae3f41187e8e\") " Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.938339 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-scripts" (OuterVolumeSpecName: "scripts") pod "a4c9cd39-beac-4a57-ba33-ae3f41187e8e" (UID: "a4c9cd39-beac-4a57-ba33-ae3f41187e8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.968655 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-kube-api-access-ggtlh" (OuterVolumeSpecName: "kube-api-access-ggtlh") pod "a4c9cd39-beac-4a57-ba33-ae3f41187e8e" (UID: "a4c9cd39-beac-4a57-ba33-ae3f41187e8e"). InnerVolumeSpecName "kube-api-access-ggtlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.968768 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-config-data" (OuterVolumeSpecName: "config-data") pod "a4c9cd39-beac-4a57-ba33-ae3f41187e8e" (UID: "a4c9cd39-beac-4a57-ba33-ae3f41187e8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:07 crc kubenswrapper[4492]: I0220 06:59:07.997686 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4c9cd39-beac-4a57-ba33-ae3f41187e8e" (UID: "a4c9cd39-beac-4a57-ba33-ae3f41187e8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.034338 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.034365 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.034376 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.034384 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggtlh\" (UniqueName: \"kubernetes.io/projected/a4c9cd39-beac-4a57-ba33-ae3f41187e8e-kube-api-access-ggtlh\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.058086 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.130223 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7558f88fd7-hk2tb"] Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.130576 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" podUID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerName="dnsmasq-dns" containerID="cri-o://e704c453c7f3c513c689ecd8855fb2b74069fbf0a5f19a792326dd662dbde1d4" gracePeriod=10 Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.431225 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerStarted","Data":"925da4d093560a26fa5667773e50ef4558d6c5ac726a12d8d33d09ad956696ba"} Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.438902 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" event={"ID":"a4c9cd39-beac-4a57-ba33-ae3f41187e8e","Type":"ContainerDied","Data":"580434c0ce39142ac30f20526ebcd65523cbfbdec4c8598f86ed25eeaeabed3b"} Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.438972 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="580434c0ce39142ac30f20526ebcd65523cbfbdec4c8598f86ed25eeaeabed3b" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.439092 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-rj7nh" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.458052 4492 generic.go:334] "Generic (PLEG): container finished" podID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerID="e704c453c7f3c513c689ecd8855fb2b74069fbf0a5f19a792326dd662dbde1d4" exitCode=0 Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.458626 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" event={"ID":"aae0abff-d540-481e-9b6d-b3e745c89db2","Type":"ContainerDied","Data":"e704c453c7f3c513c689ecd8855fb2b74069fbf0a5f19a792326dd662dbde1d4"} Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.553037 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 06:59:08 crc kubenswrapper[4492]: E0220 06:59:08.553606 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c9cd39-beac-4a57-ba33-ae3f41187e8e" containerName="nova-cell1-conductor-db-sync" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.553619 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c9cd39-beac-4a57-ba33-ae3f41187e8e" containerName="nova-cell1-conductor-db-sync" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.553787 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4c9cd39-beac-4a57-ba33-ae3f41187e8e" containerName="nova-cell1-conductor-db-sync" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.554528 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.558707 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0914d-a52b-47b6-aeed-4c14979e3afd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.558794 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0914d-a52b-47b6-aeed-4c14979e3afd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.558971 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7nf5\" (UniqueName: \"kubernetes.io/projected/eca0914d-a52b-47b6-aeed-4c14979e3afd-kube-api-access-h7nf5\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.560268 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.623946 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.625223 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.645570 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.645883 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.663275 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0914d-a52b-47b6-aeed-4c14979e3afd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.663327 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0914d-a52b-47b6-aeed-4c14979e3afd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.663423 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7nf5\" (UniqueName: \"kubernetes.io/projected/eca0914d-a52b-47b6-aeed-4c14979e3afd-kube-api-access-h7nf5\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.682778 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eca0914d-a52b-47b6-aeed-4c14979e3afd-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.698997 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7nf5\" (UniqueName: \"kubernetes.io/projected/eca0914d-a52b-47b6-aeed-4c14979e3afd-kube-api-access-h7nf5\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.699494 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eca0914d-a52b-47b6-aeed-4c14979e3afd-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eca0914d-a52b-47b6-aeed-4c14979e3afd\") " pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.780663 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.869975 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-swift-storage-0\") pod \"aae0abff-d540-481e-9b6d-b3e745c89db2\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.870365 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hjns\" (UniqueName: \"kubernetes.io/projected/aae0abff-d540-481e-9b6d-b3e745c89db2-kube-api-access-2hjns\") pod \"aae0abff-d540-481e-9b6d-b3e745c89db2\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.870525 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-svc\") pod \"aae0abff-d540-481e-9b6d-b3e745c89db2\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.870688 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-sb\") pod \"aae0abff-d540-481e-9b6d-b3e745c89db2\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.870753 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-nb\") pod \"aae0abff-d540-481e-9b6d-b3e745c89db2\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.870844 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-config\") pod \"aae0abff-d540-481e-9b6d-b3e745c89db2\" (UID: \"aae0abff-d540-481e-9b6d-b3e745c89db2\") " Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.881707 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae0abff-d540-481e-9b6d-b3e745c89db2-kube-api-access-2hjns" (OuterVolumeSpecName: "kube-api-access-2hjns") pod "aae0abff-d540-481e-9b6d-b3e745c89db2" (UID: "aae0abff-d540-481e-9b6d-b3e745c89db2"). InnerVolumeSpecName "kube-api-access-2hjns". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.939189 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.975900 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hjns\" (UniqueName: \"kubernetes.io/projected/aae0abff-d540-481e-9b6d-b3e745c89db2-kube-api-access-2hjns\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:08 crc kubenswrapper[4492]: I0220 06:59:08.984557 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.008591 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aae0abff-d540-481e-9b6d-b3e745c89db2" (UID: "aae0abff-d540-481e-9b6d-b3e745c89db2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.021627 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-config" (OuterVolumeSpecName: "config") pod "aae0abff-d540-481e-9b6d-b3e745c89db2" (UID: "aae0abff-d540-481e-9b6d-b3e745c89db2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.031382 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aae0abff-d540-481e-9b6d-b3e745c89db2" (UID: "aae0abff-d540-481e-9b6d-b3e745c89db2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.033977 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aae0abff-d540-481e-9b6d-b3e745c89db2" (UID: "aae0abff-d540-481e-9b6d-b3e745c89db2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.035612 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aae0abff-d540-481e-9b6d-b3e745c89db2" (UID: "aae0abff-d540-481e-9b6d-b3e745c89db2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.076777 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-scripts\") pod \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.076824 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-config-data\") pod \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.076938 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-combined-ca-bundle\") pod \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.077005 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6lll\" (UniqueName: \"kubernetes.io/projected/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-kube-api-access-n6lll\") pod \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\" (UID: \"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a\") " Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.077278 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.077295 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.077304 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.077315 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.077323 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aae0abff-d540-481e-9b6d-b3e745c89db2-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.080833 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-kube-api-access-n6lll" (OuterVolumeSpecName: "kube-api-access-n6lll") pod "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" (UID: "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a"). InnerVolumeSpecName "kube-api-access-n6lll". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.089580 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-scripts" (OuterVolumeSpecName: "scripts") pod "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" (UID: "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.107650 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-config-data" (OuterVolumeSpecName: "config-data") pod "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" (UID: "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.121592 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" (UID: "f0a2bbcf-c2d7-47b7-873d-4b93d85c963a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.182141 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.182165 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6lll\" (UniqueName: \"kubernetes.io/projected/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-kube-api-access-n6lll\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.182177 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.182186 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.312165 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.312224 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.467995 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" event={"ID":"aae0abff-d540-481e-9b6d-b3e745c89db2","Type":"ContainerDied","Data":"a44c83098f25b719e10d5f8cd22b596be017689372747c75ce0d2b9435fc183d"} Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.468335 4492 scope.go:117] "RemoveContainer" containerID="e704c453c7f3c513c689ecd8855fb2b74069fbf0a5f19a792326dd662dbde1d4" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.468012 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7558f88fd7-hk2tb" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.470796 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-bt5pz" event={"ID":"f0a2bbcf-c2d7-47b7-873d-4b93d85c963a","Type":"ContainerDied","Data":"bfa79cfbdd98067149873aa0ca00100152eb637b7d7d602a1f377bc12dc8d39c"} Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.470835 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-bt5pz" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.470852 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfa79cfbdd98067149873aa0ca00100152eb637b7d7d602a1f377bc12dc8d39c" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.497694 4492 scope.go:117] "RemoveContainer" containerID="ae1beeca154e35d9335bc9f4eda8f838567b3fc60fbd3ae3445bbc7b6ef851bc" Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.517822 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7558f88fd7-hk2tb"] Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.582251 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7558f88fd7-hk2tb"] Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.585155 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.618093 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.618329 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-log" containerID="cri-o://bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622" gracePeriod=30 Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.618787 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-api" containerID="cri-o://340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f" gracePeriod=30 Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.637230 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.637454 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-log" containerID="cri-o://c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c" gracePeriod=30 Feb 20 06:59:09 crc kubenswrapper[4492]: I0220 06:59:09.638897 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-metadata" containerID="cri-o://c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c" gracePeriod=30 Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.046077 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.392592 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.421557 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-config-data\") pod \"e9f53582-925f-4014-a132-5ba0d831f8c9\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.421594 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f53582-925f-4014-a132-5ba0d831f8c9-logs\") pod \"e9f53582-925f-4014-a132-5ba0d831f8c9\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.421637 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ljjm\" (UniqueName: \"kubernetes.io/projected/e9f53582-925f-4014-a132-5ba0d831f8c9-kube-api-access-5ljjm\") pod \"e9f53582-925f-4014-a132-5ba0d831f8c9\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.421797 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-combined-ca-bundle\") pod \"e9f53582-925f-4014-a132-5ba0d831f8c9\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.421827 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-nova-metadata-tls-certs\") pod \"e9f53582-925f-4014-a132-5ba0d831f8c9\" (UID: \"e9f53582-925f-4014-a132-5ba0d831f8c9\") " Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.423366 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9f53582-925f-4014-a132-5ba0d831f8c9-logs" (OuterVolumeSpecName: "logs") pod "e9f53582-925f-4014-a132-5ba0d831f8c9" (UID: "e9f53582-925f-4014-a132-5ba0d831f8c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.434750 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9f53582-925f-4014-a132-5ba0d831f8c9-kube-api-access-5ljjm" (OuterVolumeSpecName: "kube-api-access-5ljjm") pod "e9f53582-925f-4014-a132-5ba0d831f8c9" (UID: "e9f53582-925f-4014-a132-5ba0d831f8c9"). InnerVolumeSpecName "kube-api-access-5ljjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.460713 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-config-data" (OuterVolumeSpecName: "config-data") pod "e9f53582-925f-4014-a132-5ba0d831f8c9" (UID: "e9f53582-925f-4014-a132-5ba0d831f8c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.520177 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9f53582-925f-4014-a132-5ba0d831f8c9" (UID: "e9f53582-925f-4014-a132-5ba0d831f8c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.525088 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.525120 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.525130 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9f53582-925f-4014-a132-5ba0d831f8c9-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.525140 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ljjm\" (UniqueName: \"kubernetes.io/projected/e9f53582-925f-4014-a132-5ba0d831f8c9-kube-api-access-5ljjm\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.538606 4492 generic.go:334] "Generic (PLEG): container finished" podID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerID="c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c" exitCode=0 Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.538659 4492 generic.go:334] "Generic (PLEG): container finished" podID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerID="c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c" exitCode=143 Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.538728 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9f53582-925f-4014-a132-5ba0d831f8c9","Type":"ContainerDied","Data":"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c"} Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.538784 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9f53582-925f-4014-a132-5ba0d831f8c9","Type":"ContainerDied","Data":"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c"} Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.538795 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9f53582-925f-4014-a132-5ba0d831f8c9","Type":"ContainerDied","Data":"207923d592afa28b960748e40202c6761ae09f2e67a7fe608360a9c0a7ac80b4"} Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.538815 4492 scope.go:117] "RemoveContainer" containerID="c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.539016 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.548219 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerStarted","Data":"4a22d992439ef4c0214a611d5dd97cb2977c744f13016b1b956b250a266cfa80"} Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.549726 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.550492 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e9f53582-925f-4014-a132-5ba0d831f8c9" (UID: "e9f53582-925f-4014-a132-5ba0d831f8c9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.557842 4492 generic.go:334] "Generic (PLEG): container finished" podID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerID="bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622" exitCode=143 Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.557898 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31b6bf8f-6008-42dc-9ebf-a28877958e40","Type":"ContainerDied","Data":"bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622"} Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.567588 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eca0914d-a52b-47b6-aeed-4c14979e3afd","Type":"ContainerStarted","Data":"5efc77f9702fb95667aec721112ca13eb2378eb9560c534f73130d74259d94f7"} Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.567765 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eca0914d-a52b-47b6-aeed-4c14979e3afd","Type":"ContainerStarted","Data":"9e2b3f31e3c6771d859deda14ea473f8ea10323647eafb523eefb4b068c141a1"} Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.567904 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.568272 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7e4a8624-84f4-44bb-b7be-7f73c168e78b" containerName="nova-scheduler-scheduler" containerID="cri-o://7a5565002e776eb16a686355c427604138461502990fdfb251804652f7a0583e" gracePeriod=30 Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.576165 4492 scope.go:117] "RemoveContainer" containerID="c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.576513 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.17494885 podStartE2EDuration="6.576496126s" podCreationTimestamp="2026-02-20 06:59:04 +0000 UTC" firstStartedPulling="2026-02-20 06:59:05.30264268 +0000 UTC m=+1102.073931658" lastFinishedPulling="2026-02-20 06:59:09.704189956 +0000 UTC m=+1106.475478934" observedRunningTime="2026-02-20 06:59:10.563382627 +0000 UTC m=+1107.334671605" watchObservedRunningTime="2026-02-20 06:59:10.576496126 +0000 UTC m=+1107.347785104" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.594619 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.594607597 podStartE2EDuration="2.594607597s" podCreationTimestamp="2026-02-20 06:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:10.581186668 +0000 UTC m=+1107.352475646" watchObservedRunningTime="2026-02-20 06:59:10.594607597 +0000 UTC m=+1107.365896575" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.608604 4492 scope.go:117] "RemoveContainer" containerID="c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c" Feb 20 06:59:10 crc kubenswrapper[4492]: E0220 06:59:10.609833 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c\": container with ID starting with c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c not found: ID does not exist" containerID="c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.609872 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c"} err="failed to get container status \"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c\": rpc error: code = NotFound desc = could not find container \"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c\": container with ID starting with c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c not found: ID does not exist" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.609894 4492 scope.go:117] "RemoveContainer" containerID="c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c" Feb 20 06:59:10 crc kubenswrapper[4492]: E0220 06:59:10.614049 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c\": container with ID starting with c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c not found: ID does not exist" containerID="c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.614076 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c"} err="failed to get container status \"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c\": rpc error: code = NotFound desc = could not find container \"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c\": container with ID starting with c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c not found: ID does not exist" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.614096 4492 scope.go:117] "RemoveContainer" containerID="c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.618042 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c"} err="failed to get container status \"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c\": rpc error: code = NotFound desc = could not find container \"c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c\": container with ID starting with c9ef2874aada26888a7801af06c13fbfcee840a27d1365400e473e1bbdec132c not found: ID does not exist" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.618111 4492 scope.go:117] "RemoveContainer" containerID="c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.621606 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c"} err="failed to get container status \"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c\": rpc error: code = NotFound desc = could not find container \"c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c\": container with ID starting with c0a260a799e109fb1695570702fb243a4d9b745fea6e53d1b6cb638d7ba9918c not found: ID does not exist" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.626125 4492 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f53582-925f-4014-a132-5ba0d831f8c9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.875126 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.893221 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.912664 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:10 crc kubenswrapper[4492]: E0220 06:59:10.913163 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" containerName="nova-manage" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913181 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" containerName="nova-manage" Feb 20 06:59:10 crc kubenswrapper[4492]: E0220 06:59:10.913196 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-metadata" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913206 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-metadata" Feb 20 06:59:10 crc kubenswrapper[4492]: E0220 06:59:10.913216 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerName="init" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913222 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerName="init" Feb 20 06:59:10 crc kubenswrapper[4492]: E0220 06:59:10.913231 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-log" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913237 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-log" Feb 20 06:59:10 crc kubenswrapper[4492]: E0220 06:59:10.913245 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerName="dnsmasq-dns" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913250 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerName="dnsmasq-dns" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913449 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-log" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913468 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" containerName="nova-manage" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913494 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae0abff-d540-481e-9b6d-b3e745c89db2" containerName="dnsmasq-dns" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.913499 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" containerName="nova-metadata-metadata" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.914485 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.917508 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.918977 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.935809 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-logs\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.935963 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.936002 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-config-data\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.936042 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9z9g\" (UniqueName: \"kubernetes.io/projected/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-kube-api-access-v9z9g\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.936230 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:10 crc kubenswrapper[4492]: I0220 06:59:10.940330 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.038491 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.038664 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-logs\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.038784 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.038820 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-config-data\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.038856 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9z9g\" (UniqueName: \"kubernetes.io/projected/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-kube-api-access-v9z9g\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.039530 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-logs\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.042050 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.043999 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-config-data\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.044441 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.073740 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9z9g\" (UniqueName: \"kubernetes.io/projected/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-kube-api-access-v9z9g\") pod \"nova-metadata-0\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.237551 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.588330 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae0abff-d540-481e-9b6d-b3e745c89db2" path="/var/lib/kubelet/pods/aae0abff-d540-481e-9b6d-b3e745c89db2/volumes" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.589348 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9f53582-925f-4014-a132-5ba0d831f8c9" path="/var/lib/kubelet/pods/e9f53582-925f-4014-a132-5ba0d831f8c9/volumes" Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.636206 4492 generic.go:334] "Generic (PLEG): container finished" podID="7e4a8624-84f4-44bb-b7be-7f73c168e78b" containerID="7a5565002e776eb16a686355c427604138461502990fdfb251804652f7a0583e" exitCode=0 Feb 20 06:59:11 crc kubenswrapper[4492]: I0220 06:59:11.637900 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e4a8624-84f4-44bb-b7be-7f73c168e78b","Type":"ContainerDied","Data":"7a5565002e776eb16a686355c427604138461502990fdfb251804652f7a0583e"} Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.288420 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.375170 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cpq7\" (UniqueName: \"kubernetes.io/projected/7e4a8624-84f4-44bb-b7be-7f73c168e78b-kube-api-access-7cpq7\") pod \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.375579 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-combined-ca-bundle\") pod \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.376618 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data\") pod \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.385610 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.389640 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e4a8624-84f4-44bb-b7be-7f73c168e78b-kube-api-access-7cpq7" (OuterVolumeSpecName: "kube-api-access-7cpq7") pod "7e4a8624-84f4-44bb-b7be-7f73c168e78b" (UID: "7e4a8624-84f4-44bb-b7be-7f73c168e78b"). InnerVolumeSpecName "kube-api-access-7cpq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.446578 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e4a8624-84f4-44bb-b7be-7f73c168e78b" (UID: "7e4a8624-84f4-44bb-b7be-7f73c168e78b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.480605 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data" (OuterVolumeSpecName: "config-data") pod "7e4a8624-84f4-44bb-b7be-7f73c168e78b" (UID: "7e4a8624-84f4-44bb-b7be-7f73c168e78b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.481125 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data\") pod \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\" (UID: \"7e4a8624-84f4-44bb-b7be-7f73c168e78b\") " Feb 20 06:59:12 crc kubenswrapper[4492]: W0220 06:59:12.481670 4492 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/7e4a8624-84f4-44bb-b7be-7f73c168e78b/volumes/kubernetes.io~secret/config-data Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.481956 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data" (OuterVolumeSpecName: "config-data") pod "7e4a8624-84f4-44bb-b7be-7f73c168e78b" (UID: "7e4a8624-84f4-44bb-b7be-7f73c168e78b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.482856 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.482889 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cpq7\" (UniqueName: \"kubernetes.io/projected/7e4a8624-84f4-44bb-b7be-7f73c168e78b-kube-api-access-7cpq7\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.482901 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4a8624-84f4-44bb-b7be-7f73c168e78b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.683690 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7e4a8624-84f4-44bb-b7be-7f73c168e78b","Type":"ContainerDied","Data":"3ad64b8880da47a00a27934a426c81a5e29c475b1163452fa8ef79b89249ad75"} Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.684039 4492 scope.go:117] "RemoveContainer" containerID="7a5565002e776eb16a686355c427604138461502990fdfb251804652f7a0583e" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.684312 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.696310 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2127cdef-bc2d-46fe-8f3e-0934df26b4d6","Type":"ContainerStarted","Data":"26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c"} Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.696437 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2127cdef-bc2d-46fe-8f3e-0934df26b4d6","Type":"ContainerStarted","Data":"6d85aebed9ac0a8b4ed28aa7e4fb96d90a049b52588342e1e3ee87c7f2b80b45"} Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.762545 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.778365 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.805273 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:12 crc kubenswrapper[4492]: E0220 06:59:12.806353 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e4a8624-84f4-44bb-b7be-7f73c168e78b" containerName="nova-scheduler-scheduler" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.806380 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e4a8624-84f4-44bb-b7be-7f73c168e78b" containerName="nova-scheduler-scheduler" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.806670 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e4a8624-84f4-44bb-b7be-7f73c168e78b" containerName="nova-scheduler-scheduler" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.808318 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.816718 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.818055 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.895367 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjht9\" (UniqueName: \"kubernetes.io/projected/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-kube-api-access-jjht9\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.895426 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.895554 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-config-data\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.997074 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjht9\" (UniqueName: \"kubernetes.io/projected/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-kube-api-access-jjht9\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.997143 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:12 crc kubenswrapper[4492]: I0220 06:59:12.997276 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-config-data\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.001575 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-config-data\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.002006 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.013364 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjht9\" (UniqueName: \"kubernetes.io/projected/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-kube-api-access-jjht9\") pod \"nova-scheduler-0\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.137440 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.569470 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e4a8624-84f4-44bb-b7be-7f73c168e78b" path="/var/lib/kubelet/pods/7e4a8624-84f4-44bb-b7be-7f73c168e78b/volumes" Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.612438 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.729151 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2127cdef-bc2d-46fe-8f3e-0934df26b4d6","Type":"ContainerStarted","Data":"d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2"} Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.737914 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"62fd389b-b2b1-4f40-b3cc-85be0c5c4650","Type":"ContainerStarted","Data":"8909585513ea7feed0a0e725ce71c5cccf61017ac5e3aae3a4a47fc3e42ad3e2"} Feb 20 06:59:13 crc kubenswrapper[4492]: I0220 06:59:13.747516 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.747498962 podStartE2EDuration="3.747498962s" podCreationTimestamp="2026-02-20 06:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:13.745214366 +0000 UTC m=+1110.516503334" watchObservedRunningTime="2026-02-20 06:59:13.747498962 +0000 UTC m=+1110.518787940" Feb 20 06:59:14 crc kubenswrapper[4492]: I0220 06:59:14.753317 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"62fd389b-b2b1-4f40-b3cc-85be0c5c4650","Type":"ContainerStarted","Data":"6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2"} Feb 20 06:59:14 crc kubenswrapper[4492]: I0220 06:59:14.775491 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7754516110000003 podStartE2EDuration="2.775451611s" podCreationTimestamp="2026-02-20 06:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:14.76762629 +0000 UTC m=+1111.538915258" watchObservedRunningTime="2026-02-20 06:59:14.775451611 +0000 UTC m=+1111.546740579" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.509042 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.665433 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-combined-ca-bundle\") pod \"31b6bf8f-6008-42dc-9ebf-a28877958e40\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.665715 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-config-data\") pod \"31b6bf8f-6008-42dc-9ebf-a28877958e40\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.665862 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wptlh\" (UniqueName: \"kubernetes.io/projected/31b6bf8f-6008-42dc-9ebf-a28877958e40-kube-api-access-wptlh\") pod \"31b6bf8f-6008-42dc-9ebf-a28877958e40\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.666716 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b6bf8f-6008-42dc-9ebf-a28877958e40-logs\") pod \"31b6bf8f-6008-42dc-9ebf-a28877958e40\" (UID: \"31b6bf8f-6008-42dc-9ebf-a28877958e40\") " Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.667067 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31b6bf8f-6008-42dc-9ebf-a28877958e40-logs" (OuterVolumeSpecName: "logs") pod "31b6bf8f-6008-42dc-9ebf-a28877958e40" (UID: "31b6bf8f-6008-42dc-9ebf-a28877958e40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.669877 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/31b6bf8f-6008-42dc-9ebf-a28877958e40-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.673112 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31b6bf8f-6008-42dc-9ebf-a28877958e40-kube-api-access-wptlh" (OuterVolumeSpecName: "kube-api-access-wptlh") pod "31b6bf8f-6008-42dc-9ebf-a28877958e40" (UID: "31b6bf8f-6008-42dc-9ebf-a28877958e40"). InnerVolumeSpecName "kube-api-access-wptlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.694714 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-config-data" (OuterVolumeSpecName: "config-data") pod "31b6bf8f-6008-42dc-9ebf-a28877958e40" (UID: "31b6bf8f-6008-42dc-9ebf-a28877958e40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.696562 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31b6bf8f-6008-42dc-9ebf-a28877958e40" (UID: "31b6bf8f-6008-42dc-9ebf-a28877958e40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.770403 4492 generic.go:334] "Generic (PLEG): container finished" podID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerID="340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f" exitCode=0 Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.770689 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.771103 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31b6bf8f-6008-42dc-9ebf-a28877958e40","Type":"ContainerDied","Data":"340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f"} Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.771178 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"31b6bf8f-6008-42dc-9ebf-a28877958e40","Type":"ContainerDied","Data":"0970aef915bd8f1d85ccaa0912930ae3e6312b2a6a850dab94e2eceea1de2a48"} Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.771206 4492 scope.go:117] "RemoveContainer" containerID="340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.777344 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.778325 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31b6bf8f-6008-42dc-9ebf-a28877958e40-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.778360 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wptlh\" (UniqueName: \"kubernetes.io/projected/31b6bf8f-6008-42dc-9ebf-a28877958e40-kube-api-access-wptlh\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.791159 4492 scope.go:117] "RemoveContainer" containerID="bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.816706 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.820284 4492 scope.go:117] "RemoveContainer" containerID="340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f" Feb 20 06:59:15 crc kubenswrapper[4492]: E0220 06:59:15.820790 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f\": container with ID starting with 340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f not found: ID does not exist" containerID="340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.820826 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f"} err="failed to get container status \"340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f\": rpc error: code = NotFound desc = could not find container \"340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f\": container with ID starting with 340244d83c4bfed3bf162050d6d3f03a3678e49bf8f8e82c3a044e789c20d56f not found: ID does not exist" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.820851 4492 scope.go:117] "RemoveContainer" containerID="bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622" Feb 20 06:59:15 crc kubenswrapper[4492]: E0220 06:59:15.821069 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622\": container with ID starting with bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622 not found: ID does not exist" containerID="bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.821096 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622"} err="failed to get container status \"bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622\": rpc error: code = NotFound desc = could not find container \"bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622\": container with ID starting with bd83bcbc641d54a6898678ad2073128857ad9f4d946555fce68cabccf2903622 not found: ID does not exist" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.841320 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.852142 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:15 crc kubenswrapper[4492]: E0220 06:59:15.852804 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-api" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.852871 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-api" Feb 20 06:59:15 crc kubenswrapper[4492]: E0220 06:59:15.852948 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-log" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.852992 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-log" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.853286 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-log" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.853347 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" containerName="nova-api-api" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.854649 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.865046 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.869945 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.982558 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-logs\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.983213 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.983272 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-config-data\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:15 crc kubenswrapper[4492]: I0220 06:59:15.983299 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5qmf\" (UniqueName: \"kubernetes.io/projected/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-kube-api-access-g5qmf\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.087002 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-logs\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.087158 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.087216 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-config-data\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.087240 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5qmf\" (UniqueName: \"kubernetes.io/projected/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-kube-api-access-g5qmf\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.087533 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-logs\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.092026 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.093560 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-config-data\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.105111 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5qmf\" (UniqueName: \"kubernetes.io/projected/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-kube-api-access-g5qmf\") pod \"nova-api-0\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.182759 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.240268 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.240418 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.594427 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.783732 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff","Type":"ContainerStarted","Data":"5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22"} Feb 20 06:59:16 crc kubenswrapper[4492]: I0220 06:59:16.783802 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff","Type":"ContainerStarted","Data":"28f4a6e1f145e9c1fd823f956f94fecb6385de57b4264bf8738002f052c0db32"} Feb 20 06:59:17 crc kubenswrapper[4492]: I0220 06:59:17.569038 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31b6bf8f-6008-42dc-9ebf-a28877958e40" path="/var/lib/kubelet/pods/31b6bf8f-6008-42dc-9ebf-a28877958e40/volumes" Feb 20 06:59:17 crc kubenswrapper[4492]: I0220 06:59:17.797762 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff","Type":"ContainerStarted","Data":"b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b"} Feb 20 06:59:17 crc kubenswrapper[4492]: I0220 06:59:17.822222 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.822195176 podStartE2EDuration="2.822195176s" podCreationTimestamp="2026-02-20 06:59:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:17.816304903 +0000 UTC m=+1114.587593880" watchObservedRunningTime="2026-02-20 06:59:17.822195176 +0000 UTC m=+1114.593484143" Feb 20 06:59:18 crc kubenswrapper[4492]: I0220 06:59:18.138023 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 06:59:18 crc kubenswrapper[4492]: I0220 06:59:18.962625 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 20 06:59:21 crc kubenswrapper[4492]: I0220 06:59:21.240072 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 06:59:21 crc kubenswrapper[4492]: I0220 06:59:21.241205 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 06:59:22 crc kubenswrapper[4492]: I0220 06:59:22.258597 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 06:59:22 crc kubenswrapper[4492]: I0220 06:59:22.258606 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 06:59:23 crc kubenswrapper[4492]: I0220 06:59:23.138525 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 20 06:59:23 crc kubenswrapper[4492]: I0220 06:59:23.164023 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 20 06:59:23 crc kubenswrapper[4492]: I0220 06:59:23.886135 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 20 06:59:26 crc kubenswrapper[4492]: I0220 06:59:26.183638 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 06:59:26 crc kubenswrapper[4492]: I0220 06:59:26.183955 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 06:59:27 crc kubenswrapper[4492]: I0220 06:59:27.267257 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.212:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 06:59:27 crc kubenswrapper[4492]: I0220 06:59:27.267599 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.212:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 20 06:59:31 crc kubenswrapper[4492]: I0220 06:59:31.244314 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 06:59:31 crc kubenswrapper[4492]: I0220 06:59:31.247760 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 06:59:31 crc kubenswrapper[4492]: I0220 06:59:31.252418 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 06:59:31 crc kubenswrapper[4492]: I0220 06:59:31.967898 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.658100 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.697848 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-combined-ca-bundle\") pod \"fdf044a0-70f6-4682-a663-57b29bc74a17\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.697901 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-config-data\") pod \"fdf044a0-70f6-4682-a663-57b29bc74a17\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.698045 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwfc4\" (UniqueName: \"kubernetes.io/projected/fdf044a0-70f6-4682-a663-57b29bc74a17-kube-api-access-wwfc4\") pod \"fdf044a0-70f6-4682-a663-57b29bc74a17\" (UID: \"fdf044a0-70f6-4682-a663-57b29bc74a17\") " Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.703055 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdf044a0-70f6-4682-a663-57b29bc74a17-kube-api-access-wwfc4" (OuterVolumeSpecName: "kube-api-access-wwfc4") pod "fdf044a0-70f6-4682-a663-57b29bc74a17" (UID: "fdf044a0-70f6-4682-a663-57b29bc74a17"). InnerVolumeSpecName "kube-api-access-wwfc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.719933 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdf044a0-70f6-4682-a663-57b29bc74a17" (UID: "fdf044a0-70f6-4682-a663-57b29bc74a17"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.721700 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-config-data" (OuterVolumeSpecName: "config-data") pod "fdf044a0-70f6-4682-a663-57b29bc74a17" (UID: "fdf044a0-70f6-4682-a663-57b29bc74a17"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.801304 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.801339 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwfc4\" (UniqueName: \"kubernetes.io/projected/fdf044a0-70f6-4682-a663-57b29bc74a17-kube-api-access-wwfc4\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.801357 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdf044a0-70f6-4682-a663-57b29bc74a17-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:34 crc kubenswrapper[4492]: I0220 06:59:34.814313 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.004116 4492 generic.go:334] "Generic (PLEG): container finished" podID="fdf044a0-70f6-4682-a663-57b29bc74a17" containerID="51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a" exitCode=137 Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.004184 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fdf044a0-70f6-4682-a663-57b29bc74a17","Type":"ContainerDied","Data":"51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a"} Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.004229 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fdf044a0-70f6-4682-a663-57b29bc74a17","Type":"ContainerDied","Data":"467634700eb2f37c2fdc9f41873d7d0c3a28e9eee6deac89065d1460d27c1bf3"} Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.004258 4492 scope.go:117] "RemoveContainer" containerID="51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.004592 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.045585 4492 scope.go:117] "RemoveContainer" containerID="51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a" Feb 20 06:59:35 crc kubenswrapper[4492]: E0220 06:59:35.052247 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a\": container with ID starting with 51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a not found: ID does not exist" containerID="51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.052303 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a"} err="failed to get container status \"51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a\": rpc error: code = NotFound desc = could not find container \"51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a\": container with ID starting with 51b4366149a445905f6f1c40941d7c8086f96db39aaffc6c99d2f2284a473f0a not found: ID does not exist" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.056597 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.080642 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.101198 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:59:35 crc kubenswrapper[4492]: E0220 06:59:35.101950 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdf044a0-70f6-4682-a663-57b29bc74a17" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.101974 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdf044a0-70f6-4682-a663-57b29bc74a17" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.102260 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdf044a0-70f6-4682-a663-57b29bc74a17" containerName="nova-cell1-novncproxy-novncproxy" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.103302 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.110798 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.111203 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.112102 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.115729 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.115986 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.116990 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.117454 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zknt\" (UniqueName: \"kubernetes.io/projected/e720a928-304b-4ff4-9103-40ec152eaf60-kube-api-access-5zknt\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.117620 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.117863 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.220323 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.220576 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zknt\" (UniqueName: \"kubernetes.io/projected/e720a928-304b-4ff4-9103-40ec152eaf60-kube-api-access-5zknt\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.220665 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.220817 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.220886 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.228102 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.228705 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.228850 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.228887 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e720a928-304b-4ff4-9103-40ec152eaf60-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.237721 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zknt\" (UniqueName: \"kubernetes.io/projected/e720a928-304b-4ff4-9103-40ec152eaf60-kube-api-access-5zknt\") pod \"nova-cell1-novncproxy-0\" (UID: \"e720a928-304b-4ff4-9103-40ec152eaf60\") " pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.430622 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.571648 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdf044a0-70f6-4682-a663-57b29bc74a17" path="/var/lib/kubelet/pods/fdf044a0-70f6-4682-a663-57b29bc74a17/volumes" Feb 20 06:59:35 crc kubenswrapper[4492]: I0220 06:59:35.881790 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 20 06:59:36 crc kubenswrapper[4492]: I0220 06:59:36.017859 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e720a928-304b-4ff4-9103-40ec152eaf60","Type":"ContainerStarted","Data":"a5a17561b530c3416ca31d23dc3f349c675d220b399ac616f0b6fcec0a9d591a"} Feb 20 06:59:36 crc kubenswrapper[4492]: I0220 06:59:36.186596 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 06:59:36 crc kubenswrapper[4492]: I0220 06:59:36.187035 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 06:59:36 crc kubenswrapper[4492]: I0220 06:59:36.187118 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 06:59:36 crc kubenswrapper[4492]: I0220 06:59:36.189048 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.028733 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e720a928-304b-4ff4-9103-40ec152eaf60","Type":"ContainerStarted","Data":"00ebb5aeed9e8014b2cdf18aaa91fe4ac1e224ba2d12342e87142c7a68bdde1e"} Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.028821 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.037537 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.047209 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.047191784 podStartE2EDuration="2.047191784s" podCreationTimestamp="2026-02-20 06:59:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:37.040620097 +0000 UTC m=+1133.811909075" watchObservedRunningTime="2026-02-20 06:59:37.047191784 +0000 UTC m=+1133.818480762" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.273540 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65d8ddddc7-m6b6t"] Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.275394 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.280640 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65d8ddddc7-m6b6t"] Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.468392 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-sb\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.468436 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-swift-storage-0\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.468516 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-config\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.468554 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-svc\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.468584 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-nb\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.468640 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxq6k\" (UniqueName: \"kubernetes.io/projected/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-kube-api-access-cxq6k\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.571678 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-nb\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.571796 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxq6k\" (UniqueName: \"kubernetes.io/projected/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-kube-api-access-cxq6k\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.571928 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-sb\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.571952 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-swift-storage-0\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.572027 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-config\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.572073 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-svc\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.573218 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-svc\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.573272 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-nb\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.573362 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-swift-storage-0\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.573934 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-config\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.574065 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-sb\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.590158 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxq6k\" (UniqueName: \"kubernetes.io/projected/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-kube-api-access-cxq6k\") pod \"dnsmasq-dns-65d8ddddc7-m6b6t\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:37 crc kubenswrapper[4492]: I0220 06:59:37.602027 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:38 crc kubenswrapper[4492]: I0220 06:59:38.080792 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65d8ddddc7-m6b6t"] Feb 20 06:59:39 crc kubenswrapper[4492]: I0220 06:59:39.060441 4492 generic.go:334] "Generic (PLEG): container finished" podID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerID="e1e5e6202df7c982b25cf3e0d1b04f0cfe3aab232c603f22b823ef1e94cf01cd" exitCode=0 Feb 20 06:59:39 crc kubenswrapper[4492]: I0220 06:59:39.060598 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" event={"ID":"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372","Type":"ContainerDied","Data":"e1e5e6202df7c982b25cf3e0d1b04f0cfe3aab232c603f22b823ef1e94cf01cd"} Feb 20 06:59:39 crc kubenswrapper[4492]: I0220 06:59:39.061132 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" event={"ID":"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372","Type":"ContainerStarted","Data":"fd36cc307bc99820412a68a6b2e7521260fea98cfa49a7e66d4241f1626d6a9d"} Feb 20 06:59:39 crc kubenswrapper[4492]: I0220 06:59:39.311821 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 06:59:39 crc kubenswrapper[4492]: I0220 06:59:39.311895 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.073123 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" event={"ID":"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372","Type":"ContainerStarted","Data":"e1d88bafc59d2a7026a983fdf0fbaf0d489bae206d1f432191e15b7c691d0d26"} Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.073347 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.094304 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" podStartSLOduration=3.094285829 podStartE2EDuration="3.094285829s" podCreationTimestamp="2026-02-20 06:59:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:40.088353767 +0000 UTC m=+1136.859642745" watchObservedRunningTime="2026-02-20 06:59:40.094285829 +0000 UTC m=+1136.865574808" Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.236850 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.237098 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-log" containerID="cri-o://5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22" gracePeriod=30 Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.237241 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-api" containerID="cri-o://b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b" gracePeriod=30 Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.344978 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.346228 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-central-agent" containerID="cri-o://867d954a3a55c27b9f5776f996866986d5d8bbe2fb428dc7d5317e2681b797e2" gracePeriod=30 Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.346338 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="sg-core" containerID="cri-o://925da4d093560a26fa5667773e50ef4558d6c5ac726a12d8d33d09ad956696ba" gracePeriod=30 Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.346286 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-notification-agent" containerID="cri-o://155c202ecab9200b503f2a5d96bc709966f2357f1dca077be158f8c3a783fcae" gracePeriod=30 Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.346266 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="proxy-httpd" containerID="cri-o://4a22d992439ef4c0214a611d5dd97cb2977c744f13016b1b956b250a266cfa80" gracePeriod=30 Feb 20 06:59:40 crc kubenswrapper[4492]: I0220 06:59:40.431127 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.086024 4492 generic.go:334] "Generic (PLEG): container finished" podID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerID="5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22" exitCode=143 Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.086100 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff","Type":"ContainerDied","Data":"5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22"} Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.088625 4492 generic.go:334] "Generic (PLEG): container finished" podID="a995bed7-d939-4344-bac8-b506a197e28f" containerID="4a22d992439ef4c0214a611d5dd97cb2977c744f13016b1b956b250a266cfa80" exitCode=0 Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.088655 4492 generic.go:334] "Generic (PLEG): container finished" podID="a995bed7-d939-4344-bac8-b506a197e28f" containerID="925da4d093560a26fa5667773e50ef4558d6c5ac726a12d8d33d09ad956696ba" exitCode=2 Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.088663 4492 generic.go:334] "Generic (PLEG): container finished" podID="a995bed7-d939-4344-bac8-b506a197e28f" containerID="867d954a3a55c27b9f5776f996866986d5d8bbe2fb428dc7d5317e2681b797e2" exitCode=0 Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.088669 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerDied","Data":"4a22d992439ef4c0214a611d5dd97cb2977c744f13016b1b956b250a266cfa80"} Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.088716 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerDied","Data":"925da4d093560a26fa5667773e50ef4558d6c5ac726a12d8d33d09ad956696ba"} Feb 20 06:59:41 crc kubenswrapper[4492]: I0220 06:59:41.088730 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerDied","Data":"867d954a3a55c27b9f5776f996866986d5d8bbe2fb428dc7d5317e2681b797e2"} Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.797566 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.840878 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-combined-ca-bundle\") pod \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.840969 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5qmf\" (UniqueName: \"kubernetes.io/projected/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-kube-api-access-g5qmf\") pod \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.841075 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-config-data\") pod \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.841200 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-logs\") pod \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\" (UID: \"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff\") " Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.842312 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-logs" (OuterVolumeSpecName: "logs") pod "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" (UID: "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.858200 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-kube-api-access-g5qmf" (OuterVolumeSpecName: "kube-api-access-g5qmf") pod "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" (UID: "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff"). InnerVolumeSpecName "kube-api-access-g5qmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.871568 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" (UID: "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.884785 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.885020 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="03261a48-3a36-4223-b088-8d13030bdf81" containerName="kube-state-metrics" containerID="cri-o://d490c95ea1a9fe69fa81243207288d99ce3ad103fb59be07ef869a15d71816b7" gracePeriod=30 Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.956640 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-config-data" (OuterVolumeSpecName: "config-data") pod "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" (UID: "5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.959575 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.959604 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.959641 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5qmf\" (UniqueName: \"kubernetes.io/projected/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-kube-api-access-g5qmf\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:43 crc kubenswrapper[4492]: I0220 06:59:43.959654 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.130638 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03261a48-3a36-4223-b088-8d13030bdf81","Type":"ContainerDied","Data":"d490c95ea1a9fe69fa81243207288d99ce3ad103fb59be07ef869a15d71816b7"} Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.130717 4492 generic.go:334] "Generic (PLEG): container finished" podID="03261a48-3a36-4223-b088-8d13030bdf81" containerID="d490c95ea1a9fe69fa81243207288d99ce3ad103fb59be07ef869a15d71816b7" exitCode=2 Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.140901 4492 generic.go:334] "Generic (PLEG): container finished" podID="a995bed7-d939-4344-bac8-b506a197e28f" containerID="155c202ecab9200b503f2a5d96bc709966f2357f1dca077be158f8c3a783fcae" exitCode=0 Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.140959 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerDied","Data":"155c202ecab9200b503f2a5d96bc709966f2357f1dca077be158f8c3a783fcae"} Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.146139 4492 generic.go:334] "Generic (PLEG): container finished" podID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerID="b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b" exitCode=0 Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.146171 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff","Type":"ContainerDied","Data":"b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b"} Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.146197 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff","Type":"ContainerDied","Data":"28f4a6e1f145e9c1fd823f956f94fecb6385de57b4264bf8738002f052c0db32"} Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.146219 4492 scope.go:117] "RemoveContainer" containerID="b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.146400 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.199582 4492 scope.go:117] "RemoveContainer" containerID="5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.216601 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.237466 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.241688 4492 scope.go:117] "RemoveContainer" containerID="b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b" Feb 20 06:59:44 crc kubenswrapper[4492]: E0220 06:59:44.243263 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b\": container with ID starting with b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b not found: ID does not exist" containerID="b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.243304 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b"} err="failed to get container status \"b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b\": rpc error: code = NotFound desc = could not find container \"b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b\": container with ID starting with b919d153c0d8148aed07812f92adef8a265067a0c44e53e55595acd05015967b not found: ID does not exist" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.243330 4492 scope.go:117] "RemoveContainer" containerID="5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22" Feb 20 06:59:44 crc kubenswrapper[4492]: E0220 06:59:44.243726 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22\": container with ID starting with 5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22 not found: ID does not exist" containerID="5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.243749 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22"} err="failed to get container status \"5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22\": rpc error: code = NotFound desc = could not find container \"5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22\": container with ID starting with 5be48d5be657ce7dacfecaab4d3d116ea2b39d6407621b0adf206653f7215a22 not found: ID does not exist" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.257721 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:44 crc kubenswrapper[4492]: E0220 06:59:44.258190 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-log" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.258208 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-log" Feb 20 06:59:44 crc kubenswrapper[4492]: E0220 06:59:44.258227 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-api" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.258234 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-api" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.258426 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-log" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.258453 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" containerName="nova-api-api" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.259507 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.264823 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.264948 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.266314 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.267750 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.339228 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.375225 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vlzg\" (UniqueName: \"kubernetes.io/projected/ca3d78d6-072d-4a79-9325-aa054728bca4-kube-api-access-8vlzg\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.375420 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.375682 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-public-tls-certs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.375711 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.375840 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-config-data\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.375875 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca3d78d6-072d-4a79-9325-aa054728bca4-logs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.477662 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjncw\" (UniqueName: \"kubernetes.io/projected/03261a48-3a36-4223-b088-8d13030bdf81-kube-api-access-vjncw\") pod \"03261a48-3a36-4223-b088-8d13030bdf81\" (UID: \"03261a48-3a36-4223-b088-8d13030bdf81\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.478245 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-config-data\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.478296 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca3d78d6-072d-4a79-9325-aa054728bca4-logs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.478454 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vlzg\" (UniqueName: \"kubernetes.io/projected/ca3d78d6-072d-4a79-9325-aa054728bca4-kube-api-access-8vlzg\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.478586 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.478639 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-public-tls-certs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.478661 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.479894 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca3d78d6-072d-4a79-9325-aa054728bca4-logs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.486176 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03261a48-3a36-4223-b088-8d13030bdf81-kube-api-access-vjncw" (OuterVolumeSpecName: "kube-api-access-vjncw") pod "03261a48-3a36-4223-b088-8d13030bdf81" (UID: "03261a48-3a36-4223-b088-8d13030bdf81"). InnerVolumeSpecName "kube-api-access-vjncw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.487127 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.490811 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-public-tls-certs\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.491343 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.494359 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-config-data\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.500068 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vlzg\" (UniqueName: \"kubernetes.io/projected/ca3d78d6-072d-4a79-9325-aa054728bca4-kube-api-access-8vlzg\") pod \"nova-api-0\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.552625 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.581462 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.589008 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjncw\" (UniqueName: \"kubernetes.io/projected/03261a48-3a36-4223-b088-8d13030bdf81-kube-api-access-vjncw\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.692173 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-run-httpd\") pod \"a995bed7-d939-4344-bac8-b506a197e28f\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.692439 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-scripts\") pod \"a995bed7-d939-4344-bac8-b506a197e28f\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.692489 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4c6x\" (UniqueName: \"kubernetes.io/projected/a995bed7-d939-4344-bac8-b506a197e28f-kube-api-access-k4c6x\") pod \"a995bed7-d939-4344-bac8-b506a197e28f\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.692555 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-sg-core-conf-yaml\") pod \"a995bed7-d939-4344-bac8-b506a197e28f\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.692677 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-config-data\") pod \"a995bed7-d939-4344-bac8-b506a197e28f\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.692699 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-combined-ca-bundle\") pod \"a995bed7-d939-4344-bac8-b506a197e28f\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.692730 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-log-httpd\") pod \"a995bed7-d939-4344-bac8-b506a197e28f\" (UID: \"a995bed7-d939-4344-bac8-b506a197e28f\") " Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.694657 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a995bed7-d939-4344-bac8-b506a197e28f" (UID: "a995bed7-d939-4344-bac8-b506a197e28f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.695786 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a995bed7-d939-4344-bac8-b506a197e28f" (UID: "a995bed7-d939-4344-bac8-b506a197e28f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.700010 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-scripts" (OuterVolumeSpecName: "scripts") pod "a995bed7-d939-4344-bac8-b506a197e28f" (UID: "a995bed7-d939-4344-bac8-b506a197e28f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.704300 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a995bed7-d939-4344-bac8-b506a197e28f-kube-api-access-k4c6x" (OuterVolumeSpecName: "kube-api-access-k4c6x") pod "a995bed7-d939-4344-bac8-b506a197e28f" (UID: "a995bed7-d939-4344-bac8-b506a197e28f"). InnerVolumeSpecName "kube-api-access-k4c6x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.753610 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a995bed7-d939-4344-bac8-b506a197e28f" (UID: "a995bed7-d939-4344-bac8-b506a197e28f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.797024 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.797145 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.797262 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a995bed7-d939-4344-bac8-b506a197e28f-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.797314 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4c6x\" (UniqueName: \"kubernetes.io/projected/a995bed7-d939-4344-bac8-b506a197e28f-kube-api-access-k4c6x\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.797359 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.812838 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a995bed7-d939-4344-bac8-b506a197e28f" (UID: "a995bed7-d939-4344-bac8-b506a197e28f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.840593 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-config-data" (OuterVolumeSpecName: "config-data") pod "a995bed7-d939-4344-bac8-b506a197e28f" (UID: "a995bed7-d939-4344-bac8-b506a197e28f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.899535 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:44 crc kubenswrapper[4492]: I0220 06:59:44.899564 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a995bed7-d939-4344-bac8-b506a197e28f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.074174 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.155586 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.155585 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"03261a48-3a36-4223-b088-8d13030bdf81","Type":"ContainerDied","Data":"6bf7bb27cfa1bff20688e19b4fda54af6dab7e2fae9114ed754d9afbe11b4aca"} Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.155758 4492 scope.go:117] "RemoveContainer" containerID="d490c95ea1a9fe69fa81243207288d99ce3ad103fb59be07ef869a15d71816b7" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.162747 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a995bed7-d939-4344-bac8-b506a197e28f","Type":"ContainerDied","Data":"03dcb71bec1296c043fabf557ef7d12deb827d612d472e894313d06980067df5"} Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.162858 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.169584 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca3d78d6-072d-4a79-9325-aa054728bca4","Type":"ContainerStarted","Data":"274c0912c6d8ff2471472f68dc59458c71fa4a9560cbd2232878e010fa933f63"} Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.176293 4492 scope.go:117] "RemoveContainer" containerID="4a22d992439ef4c0214a611d5dd97cb2977c744f13016b1b956b250a266cfa80" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.197511 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.202714 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.205584 4492 scope.go:117] "RemoveContainer" containerID="925da4d093560a26fa5667773e50ef4558d6c5ac726a12d8d33d09ad956696ba" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.207594 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.214361 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.219838 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: E0220 06:59:45.220158 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="proxy-httpd" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220173 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="proxy-httpd" Feb 20 06:59:45 crc kubenswrapper[4492]: E0220 06:59:45.220192 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-notification-agent" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220199 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-notification-agent" Feb 20 06:59:45 crc kubenswrapper[4492]: E0220 06:59:45.220215 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03261a48-3a36-4223-b088-8d13030bdf81" containerName="kube-state-metrics" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220221 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="03261a48-3a36-4223-b088-8d13030bdf81" containerName="kube-state-metrics" Feb 20 06:59:45 crc kubenswrapper[4492]: E0220 06:59:45.220234 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="sg-core" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220241 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="sg-core" Feb 20 06:59:45 crc kubenswrapper[4492]: E0220 06:59:45.220254 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-central-agent" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220260 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-central-agent" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220393 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="03261a48-3a36-4223-b088-8d13030bdf81" containerName="kube-state-metrics" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220416 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="proxy-httpd" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220427 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="sg-core" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220435 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-notification-agent" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.220443 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a995bed7-d939-4344-bac8-b506a197e28f" containerName="ceilometer-central-agent" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.221375 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.227096 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-xm49w" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.227254 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.228138 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.237560 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.248144 4492 scope.go:117] "RemoveContainer" containerID="155c202ecab9200b503f2a5d96bc709966f2357f1dca077be158f8c3a783fcae" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.279828 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.282678 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.284237 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.289751 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.296542 4492 scope.go:117] "RemoveContainer" containerID="867d954a3a55c27b9f5776f996866986d5d8bbe2fb428dc7d5317e2681b797e2" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.305827 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.305984 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.306090 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmpc9\" (UniqueName: \"kubernetes.io/projected/97f89595-4092-4ed0-acc5-421443827a47-kube-api-access-tmpc9\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.306149 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.322892 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.409373 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmpc9\" (UniqueName: \"kubernetes.io/projected/97f89595-4092-4ed0-acc5-421443827a47-kube-api-access-tmpc9\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.409424 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.409516 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-scripts\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.409542 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.409567 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.409589 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.409893 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srw8c\" (UniqueName: \"kubernetes.io/projected/f0d824b3-b6dd-45d7-9675-0042da9a3048-kube-api-access-srw8c\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.410011 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.410074 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-config-data\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.410138 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.410162 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.414997 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.417238 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.417292 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97f89595-4092-4ed0-acc5-421443827a47-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.424975 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmpc9\" (UniqueName: \"kubernetes.io/projected/97f89595-4092-4ed0-acc5-421443827a47-kube-api-access-tmpc9\") pod \"kube-state-metrics-0\" (UID: \"97f89595-4092-4ed0-acc5-421443827a47\") " pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.430948 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.452949 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.520640 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-scripts\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.520692 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.520724 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.520748 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.520882 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srw8c\" (UniqueName: \"kubernetes.io/projected/f0d824b3-b6dd-45d7-9675-0042da9a3048-kube-api-access-srw8c\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.521014 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-config-data\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.521116 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.529085 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-log-httpd\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.529564 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-run-httpd\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.539905 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.541421 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.542013 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.549989 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-config-data\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.551596 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-scripts\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.562225 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srw8c\" (UniqueName: \"kubernetes.io/projected/f0d824b3-b6dd-45d7-9675-0042da9a3048-kube-api-access-srw8c\") pod \"ceilometer-0\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " pod="openstack/ceilometer-0" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.587938 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03261a48-3a36-4223-b088-8d13030bdf81" path="/var/lib/kubelet/pods/03261a48-3a36-4223-b088-8d13030bdf81/volumes" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.588793 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff" path="/var/lib/kubelet/pods/5e5ec1a4-0844-45c8-9b37-2276c0e1f1ff/volumes" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.590119 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a995bed7-d939-4344-bac8-b506a197e28f" path="/var/lib/kubelet/pods/a995bed7-d939-4344-bac8-b506a197e28f/volumes" Feb 20 06:59:45 crc kubenswrapper[4492]: I0220 06:59:45.597516 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.024692 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 20 06:59:46 crc kubenswrapper[4492]: W0220 06:59:46.024996 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97f89595_4092_4ed0_acc5_421443827a47.slice/crio-e0d3182b8cb214e30fe2bd74f42e6ba8ea696ed813cb10203e4bd4586abfc8c8 WatchSource:0}: Error finding container e0d3182b8cb214e30fe2bd74f42e6ba8ea696ed813cb10203e4bd4586abfc8c8: Status 404 returned error can't find the container with id e0d3182b8cb214e30fe2bd74f42e6ba8ea696ed813cb10203e4bd4586abfc8c8 Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.029158 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.141814 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.182557 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerStarted","Data":"a1f90824544d0d77c4fbb56e3be3761564b58753f1ca13169ffc8ce8ace1991d"} Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.185133 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca3d78d6-072d-4a79-9325-aa054728bca4","Type":"ContainerStarted","Data":"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136"} Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.185199 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca3d78d6-072d-4a79-9325-aa054728bca4","Type":"ContainerStarted","Data":"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341"} Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.188092 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"97f89595-4092-4ed0-acc5-421443827a47","Type":"ContainerStarted","Data":"e0d3182b8cb214e30fe2bd74f42e6ba8ea696ed813cb10203e4bd4586abfc8c8"} Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.211979 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.211960282 podStartE2EDuration="2.211960282s" podCreationTimestamp="2026-02-20 06:59:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:46.207809627 +0000 UTC m=+1142.979098605" watchObservedRunningTime="2026-02-20 06:59:46.211960282 +0000 UTC m=+1142.983249259" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.226522 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.389887 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.526173 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-slhpn"] Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.528348 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.533041 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.533202 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.547005 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-slhpn"] Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.662255 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-config-data\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.662350 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnlqv\" (UniqueName: \"kubernetes.io/projected/7ccbd7a5-60ae-4701-8365-bf880455f53e-kube-api-access-nnlqv\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.662380 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.662578 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-scripts\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.765552 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-config-data\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.765618 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnlqv\" (UniqueName: \"kubernetes.io/projected/7ccbd7a5-60ae-4701-8365-bf880455f53e-kube-api-access-nnlqv\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.765649 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.765832 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-scripts\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.772161 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-scripts\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.773688 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-config-data\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.780718 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.784065 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnlqv\" (UniqueName: \"kubernetes.io/projected/7ccbd7a5-60ae-4701-8365-bf880455f53e-kube-api-access-nnlqv\") pod \"nova-cell1-cell-mapping-slhpn\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:46 crc kubenswrapper[4492]: I0220 06:59:46.845617 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.206891 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"97f89595-4092-4ed0-acc5-421443827a47","Type":"ContainerStarted","Data":"0ed68184260c3081b56a9f17d221f8fcdd4e8f955d44597109fd2c28180ddc1f"} Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.208694 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.212747 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerStarted","Data":"0ba6213840b64a97b086278a8e58a0b5d909e64748c1fbb84985473e03d96fdc"} Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.226408 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.9285038349999999 podStartE2EDuration="2.226393665s" podCreationTimestamp="2026-02-20 06:59:45 +0000 UTC" firstStartedPulling="2026-02-20 06:59:46.027640653 +0000 UTC m=+1142.798929632" lastFinishedPulling="2026-02-20 06:59:46.325530485 +0000 UTC m=+1143.096819462" observedRunningTime="2026-02-20 06:59:47.223547992 +0000 UTC m=+1143.994836969" watchObservedRunningTime="2026-02-20 06:59:47.226393665 +0000 UTC m=+1143.997682643" Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.312371 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-slhpn"] Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.603426 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.660408 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b946fdf6f-dz5bm"] Feb 20 06:59:47 crc kubenswrapper[4492]: I0220 06:59:47.660658 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerName="dnsmasq-dns" containerID="cri-o://4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550" gracePeriod=10 Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.187261 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.239075 4492 generic.go:334] "Generic (PLEG): container finished" podID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerID="4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550" exitCode=0 Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.239186 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" event={"ID":"26c40a07-b59c-4796-a5cb-caf6d8482de2","Type":"ContainerDied","Data":"4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550"} Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.239235 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" event={"ID":"26c40a07-b59c-4796-a5cb-caf6d8482de2","Type":"ContainerDied","Data":"6c7d1e8174755ecfcf503de1b6cb3b17bf95b81e05f503f868a7e1dbe237445f"} Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.239276 4492 scope.go:117] "RemoveContainer" containerID="4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.239504 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.246294 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerStarted","Data":"b4d029bcec4c6834ecdadeb94cb29ccb139438b281e1e38aacb85375f963379f"} Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.270940 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-slhpn" event={"ID":"7ccbd7a5-60ae-4701-8365-bf880455f53e","Type":"ContainerStarted","Data":"aeaff778beb14841dc9c2652898e241fcc093e71ae51937fee55d8ed24173096"} Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.270988 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-slhpn" event={"ID":"7ccbd7a5-60ae-4701-8365-bf880455f53e","Type":"ContainerStarted","Data":"3a937d7b0f129de5f7dc1a79ece8291cc03dee618f1520349a64af8ccc4cb3fe"} Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.284748 4492 scope.go:117] "RemoveContainer" containerID="37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.304063 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-slhpn" podStartSLOduration=2.304050597 podStartE2EDuration="2.304050597s" podCreationTimestamp="2026-02-20 06:59:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:48.300949403 +0000 UTC m=+1145.072238380" watchObservedRunningTime="2026-02-20 06:59:48.304050597 +0000 UTC m=+1145.075339576" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.308050 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-swift-storage-0\") pod \"26c40a07-b59c-4796-a5cb-caf6d8482de2\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.308163 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-nb\") pod \"26c40a07-b59c-4796-a5cb-caf6d8482de2\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.308205 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-sb\") pod \"26c40a07-b59c-4796-a5cb-caf6d8482de2\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.308283 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d555\" (UniqueName: \"kubernetes.io/projected/26c40a07-b59c-4796-a5cb-caf6d8482de2-kube-api-access-5d555\") pod \"26c40a07-b59c-4796-a5cb-caf6d8482de2\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.308317 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-config\") pod \"26c40a07-b59c-4796-a5cb-caf6d8482de2\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.308353 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-svc\") pod \"26c40a07-b59c-4796-a5cb-caf6d8482de2\" (UID: \"26c40a07-b59c-4796-a5cb-caf6d8482de2\") " Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.317632 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26c40a07-b59c-4796-a5cb-caf6d8482de2-kube-api-access-5d555" (OuterVolumeSpecName: "kube-api-access-5d555") pod "26c40a07-b59c-4796-a5cb-caf6d8482de2" (UID: "26c40a07-b59c-4796-a5cb-caf6d8482de2"). InnerVolumeSpecName "kube-api-access-5d555". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.317720 4492 scope.go:117] "RemoveContainer" containerID="4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550" Feb 20 06:59:48 crc kubenswrapper[4492]: E0220 06:59:48.318161 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550\": container with ID starting with 4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550 not found: ID does not exist" containerID="4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.318192 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550"} err="failed to get container status \"4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550\": rpc error: code = NotFound desc = could not find container \"4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550\": container with ID starting with 4cfcb1f04bc32096ba58f33dc823b34f9cf162c9a8c3f5819572d90a44ec6550 not found: ID does not exist" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.318212 4492 scope.go:117] "RemoveContainer" containerID="37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e" Feb 20 06:59:48 crc kubenswrapper[4492]: E0220 06:59:48.318444 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e\": container with ID starting with 37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e not found: ID does not exist" containerID="37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.318464 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e"} err="failed to get container status \"37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e\": rpc error: code = NotFound desc = could not find container \"37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e\": container with ID starting with 37e73b853a020453309b4ad0705c75fee0df5712c8004d561142fa1e2076217e not found: ID does not exist" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.363660 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26c40a07-b59c-4796-a5cb-caf6d8482de2" (UID: "26c40a07-b59c-4796-a5cb-caf6d8482de2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.378347 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-config" (OuterVolumeSpecName: "config") pod "26c40a07-b59c-4796-a5cb-caf6d8482de2" (UID: "26c40a07-b59c-4796-a5cb-caf6d8482de2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.378860 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26c40a07-b59c-4796-a5cb-caf6d8482de2" (UID: "26c40a07-b59c-4796-a5cb-caf6d8482de2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.379737 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "26c40a07-b59c-4796-a5cb-caf6d8482de2" (UID: "26c40a07-b59c-4796-a5cb-caf6d8482de2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.384969 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26c40a07-b59c-4796-a5cb-caf6d8482de2" (UID: "26c40a07-b59c-4796-a5cb-caf6d8482de2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.412438 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.412459 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.412483 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.412492 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d555\" (UniqueName: \"kubernetes.io/projected/26c40a07-b59c-4796-a5cb-caf6d8482de2-kube-api-access-5d555\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.412500 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-config\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.412508 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26c40a07-b59c-4796-a5cb-caf6d8482de2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.564998 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b946fdf6f-dz5bm"] Feb 20 06:59:48 crc kubenswrapper[4492]: I0220 06:59:48.570200 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b946fdf6f-dz5bm"] Feb 20 06:59:49 crc kubenswrapper[4492]: I0220 06:59:49.286626 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerStarted","Data":"83d595e0abde2bdad7d59fe4d7f42eb9afbe231390e9f8d9011d47b5da26af1a"} Feb 20 06:59:49 crc kubenswrapper[4492]: I0220 06:59:49.569090 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" path="/var/lib/kubelet/pods/26c40a07-b59c-4796-a5cb-caf6d8482de2/volumes" Feb 20 06:59:51 crc kubenswrapper[4492]: I0220 06:59:51.305400 4492 generic.go:334] "Generic (PLEG): container finished" podID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerID="a14db54d07ab853296454c4c575a150a36b4f67a1586d2feed06415b205edfe1" exitCode=1 Feb 20 06:59:51 crc kubenswrapper[4492]: I0220 06:59:51.305505 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerDied","Data":"a14db54d07ab853296454c4c575a150a36b4f67a1586d2feed06415b205edfe1"} Feb 20 06:59:51 crc kubenswrapper[4492]: I0220 06:59:51.305566 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-central-agent" containerID="cri-o://0ba6213840b64a97b086278a8e58a0b5d909e64748c1fbb84985473e03d96fdc" gracePeriod=30 Feb 20 06:59:51 crc kubenswrapper[4492]: I0220 06:59:51.305581 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="sg-core" containerID="cri-o://83d595e0abde2bdad7d59fe4d7f42eb9afbe231390e9f8d9011d47b5da26af1a" gracePeriod=30 Feb 20 06:59:51 crc kubenswrapper[4492]: I0220 06:59:51.305613 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-notification-agent" containerID="cri-o://b4d029bcec4c6834ecdadeb94cb29ccb139438b281e1e38aacb85375f963379f" gracePeriod=30 Feb 20 06:59:52 crc kubenswrapper[4492]: I0220 06:59:52.325323 4492 generic.go:334] "Generic (PLEG): container finished" podID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerID="83d595e0abde2bdad7d59fe4d7f42eb9afbe231390e9f8d9011d47b5da26af1a" exitCode=2 Feb 20 06:59:52 crc kubenswrapper[4492]: I0220 06:59:52.325680 4492 generic.go:334] "Generic (PLEG): container finished" podID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerID="b4d029bcec4c6834ecdadeb94cb29ccb139438b281e1e38aacb85375f963379f" exitCode=0 Feb 20 06:59:52 crc kubenswrapper[4492]: I0220 06:59:52.325739 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerDied","Data":"83d595e0abde2bdad7d59fe4d7f42eb9afbe231390e9f8d9011d47b5da26af1a"} Feb 20 06:59:52 crc kubenswrapper[4492]: I0220 06:59:52.325789 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerDied","Data":"b4d029bcec4c6834ecdadeb94cb29ccb139438b281e1e38aacb85375f963379f"} Feb 20 06:59:52 crc kubenswrapper[4492]: I0220 06:59:52.328466 4492 generic.go:334] "Generic (PLEG): container finished" podID="7ccbd7a5-60ae-4701-8365-bf880455f53e" containerID="aeaff778beb14841dc9c2652898e241fcc093e71ae51937fee55d8ed24173096" exitCode=0 Feb 20 06:59:52 crc kubenswrapper[4492]: I0220 06:59:52.328515 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-slhpn" event={"ID":"7ccbd7a5-60ae-4701-8365-bf880455f53e","Type":"ContainerDied","Data":"aeaff778beb14841dc9c2652898e241fcc093e71ae51937fee55d8ed24173096"} Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.056662 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b946fdf6f-dz5bm" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.205:5353: i/o timeout" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.668345 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.764383 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-scripts\") pod \"7ccbd7a5-60ae-4701-8365-bf880455f53e\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.764507 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-combined-ca-bundle\") pod \"7ccbd7a5-60ae-4701-8365-bf880455f53e\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.764554 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnlqv\" (UniqueName: \"kubernetes.io/projected/7ccbd7a5-60ae-4701-8365-bf880455f53e-kube-api-access-nnlqv\") pod \"7ccbd7a5-60ae-4701-8365-bf880455f53e\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.764607 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-config-data\") pod \"7ccbd7a5-60ae-4701-8365-bf880455f53e\" (UID: \"7ccbd7a5-60ae-4701-8365-bf880455f53e\") " Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.774522 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-scripts" (OuterVolumeSpecName: "scripts") pod "7ccbd7a5-60ae-4701-8365-bf880455f53e" (UID: "7ccbd7a5-60ae-4701-8365-bf880455f53e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.774542 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ccbd7a5-60ae-4701-8365-bf880455f53e-kube-api-access-nnlqv" (OuterVolumeSpecName: "kube-api-access-nnlqv") pod "7ccbd7a5-60ae-4701-8365-bf880455f53e" (UID: "7ccbd7a5-60ae-4701-8365-bf880455f53e"). InnerVolumeSpecName "kube-api-access-nnlqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.789462 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-config-data" (OuterVolumeSpecName: "config-data") pod "7ccbd7a5-60ae-4701-8365-bf880455f53e" (UID: "7ccbd7a5-60ae-4701-8365-bf880455f53e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.790186 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ccbd7a5-60ae-4701-8365-bf880455f53e" (UID: "7ccbd7a5-60ae-4701-8365-bf880455f53e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.870416 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.870454 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnlqv\" (UniqueName: \"kubernetes.io/projected/7ccbd7a5-60ae-4701-8365-bf880455f53e-kube-api-access-nnlqv\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.870468 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:53 crc kubenswrapper[4492]: I0220 06:59:53.870499 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccbd7a5-60ae-4701-8365-bf880455f53e-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.356349 4492 generic.go:334] "Generic (PLEG): container finished" podID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerID="0ba6213840b64a97b086278a8e58a0b5d909e64748c1fbb84985473e03d96fdc" exitCode=0 Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.356431 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerDied","Data":"0ba6213840b64a97b086278a8e58a0b5d909e64748c1fbb84985473e03d96fdc"} Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.369045 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-slhpn" event={"ID":"7ccbd7a5-60ae-4701-8365-bf880455f53e","Type":"ContainerDied","Data":"3a937d7b0f129de5f7dc1a79ece8291cc03dee618f1520349a64af8ccc4cb3fe"} Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.369113 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a937d7b0f129de5f7dc1a79ece8291cc03dee618f1520349a64af8ccc4cb3fe" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.369215 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-slhpn" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.402800 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.482748 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-log-httpd\") pod \"f0d824b3-b6dd-45d7-9675-0042da9a3048\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.483169 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-sg-core-conf-yaml\") pod \"f0d824b3-b6dd-45d7-9675-0042da9a3048\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.483308 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-scripts\") pod \"f0d824b3-b6dd-45d7-9675-0042da9a3048\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.483428 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-config-data\") pod \"f0d824b3-b6dd-45d7-9675-0042da9a3048\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.483517 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-run-httpd\") pod \"f0d824b3-b6dd-45d7-9675-0042da9a3048\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.483617 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srw8c\" (UniqueName: \"kubernetes.io/projected/f0d824b3-b6dd-45d7-9675-0042da9a3048-kube-api-access-srw8c\") pod \"f0d824b3-b6dd-45d7-9675-0042da9a3048\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.483693 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-combined-ca-bundle\") pod \"f0d824b3-b6dd-45d7-9675-0042da9a3048\" (UID: \"f0d824b3-b6dd-45d7-9675-0042da9a3048\") " Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.483301 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f0d824b3-b6dd-45d7-9675-0042da9a3048" (UID: "f0d824b3-b6dd-45d7-9675-0042da9a3048"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.484023 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f0d824b3-b6dd-45d7-9675-0042da9a3048" (UID: "f0d824b3-b6dd-45d7-9675-0042da9a3048"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.484790 4492 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.484854 4492 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f0d824b3-b6dd-45d7-9675-0042da9a3048-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.494408 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-scripts" (OuterVolumeSpecName: "scripts") pod "f0d824b3-b6dd-45d7-9675-0042da9a3048" (UID: "f0d824b3-b6dd-45d7-9675-0042da9a3048"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.499778 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0d824b3-b6dd-45d7-9675-0042da9a3048-kube-api-access-srw8c" (OuterVolumeSpecName: "kube-api-access-srw8c") pod "f0d824b3-b6dd-45d7-9675-0042da9a3048" (UID: "f0d824b3-b6dd-45d7-9675-0042da9a3048"). InnerVolumeSpecName "kube-api-access-srw8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.547560 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.547923 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="62fd389b-b2b1-4f40-b3cc-85be0c5c4650" containerName="nova-scheduler-scheduler" containerID="cri-o://6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2" gracePeriod=30 Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.553080 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.553312 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-log" containerID="cri-o://3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341" gracePeriod=30 Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.554621 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-api" containerID="cri-o://e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136" gracePeriod=30 Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.563341 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.563762 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-log" containerID="cri-o://26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c" gracePeriod=30 Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.563942 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-metadata" containerID="cri-o://d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2" gracePeriod=30 Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.569733 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f0d824b3-b6dd-45d7-9675-0042da9a3048" (UID: "f0d824b3-b6dd-45d7-9675-0042da9a3048"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.586314 4492 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.586391 4492 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-scripts\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.586451 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srw8c\" (UniqueName: \"kubernetes.io/projected/f0d824b3-b6dd-45d7-9675-0042da9a3048-kube-api-access-srw8c\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.611200 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0d824b3-b6dd-45d7-9675-0042da9a3048" (UID: "f0d824b3-b6dd-45d7-9675-0042da9a3048"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.622091 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-config-data" (OuterVolumeSpecName: "config-data") pod "f0d824b3-b6dd-45d7-9675-0042da9a3048" (UID: "f0d824b3-b6dd-45d7-9675-0042da9a3048"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.687977 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:54 crc kubenswrapper[4492]: I0220 06:59:54.689289 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d824b3-b6dd-45d7-9675-0042da9a3048-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.045736 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.098733 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-config-data\") pod \"ca3d78d6-072d-4a79-9325-aa054728bca4\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.099057 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca3d78d6-072d-4a79-9325-aa054728bca4-logs\") pod \"ca3d78d6-072d-4a79-9325-aa054728bca4\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.099131 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-public-tls-certs\") pod \"ca3d78d6-072d-4a79-9325-aa054728bca4\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.099204 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-internal-tls-certs\") pod \"ca3d78d6-072d-4a79-9325-aa054728bca4\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.099269 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vlzg\" (UniqueName: \"kubernetes.io/projected/ca3d78d6-072d-4a79-9325-aa054728bca4-kube-api-access-8vlzg\") pod \"ca3d78d6-072d-4a79-9325-aa054728bca4\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.099315 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-combined-ca-bundle\") pod \"ca3d78d6-072d-4a79-9325-aa054728bca4\" (UID: \"ca3d78d6-072d-4a79-9325-aa054728bca4\") " Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.099381 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca3d78d6-072d-4a79-9325-aa054728bca4-logs" (OuterVolumeSpecName: "logs") pod "ca3d78d6-072d-4a79-9325-aa054728bca4" (UID: "ca3d78d6-072d-4a79-9325-aa054728bca4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.100242 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca3d78d6-072d-4a79-9325-aa054728bca4-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.104991 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca3d78d6-072d-4a79-9325-aa054728bca4-kube-api-access-8vlzg" (OuterVolumeSpecName: "kube-api-access-8vlzg") pod "ca3d78d6-072d-4a79-9325-aa054728bca4" (UID: "ca3d78d6-072d-4a79-9325-aa054728bca4"). InnerVolumeSpecName "kube-api-access-8vlzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.138676 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-config-data" (OuterVolumeSpecName: "config-data") pod "ca3d78d6-072d-4a79-9325-aa054728bca4" (UID: "ca3d78d6-072d-4a79-9325-aa054728bca4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.144620 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca3d78d6-072d-4a79-9325-aa054728bca4" (UID: "ca3d78d6-072d-4a79-9325-aa054728bca4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.150147 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ca3d78d6-072d-4a79-9325-aa054728bca4" (UID: "ca3d78d6-072d-4a79-9325-aa054728bca4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.165573 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ca3d78d6-072d-4a79-9325-aa054728bca4" (UID: "ca3d78d6-072d-4a79-9325-aa054728bca4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.211899 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.212368 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.212446 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vlzg\" (UniqueName: \"kubernetes.io/projected/ca3d78d6-072d-4a79-9325-aa054728bca4-kube-api-access-8vlzg\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.212545 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.212595 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3d78d6-072d-4a79-9325-aa054728bca4-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.383380 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f0d824b3-b6dd-45d7-9675-0042da9a3048","Type":"ContainerDied","Data":"a1f90824544d0d77c4fbb56e3be3761564b58753f1ca13169ffc8ce8ace1991d"} Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.383520 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.383633 4492 scope.go:117] "RemoveContainer" containerID="a14db54d07ab853296454c4c575a150a36b4f67a1586d2feed06415b205edfe1" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.386496 4492 generic.go:334] "Generic (PLEG): container finished" podID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerID="26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c" exitCode=143 Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.386609 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2127cdef-bc2d-46fe-8f3e-0934df26b4d6","Type":"ContainerDied","Data":"26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c"} Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.389566 4492 generic.go:334] "Generic (PLEG): container finished" podID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerID="e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136" exitCode=0 Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.389707 4492 generic.go:334] "Generic (PLEG): container finished" podID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerID="3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341" exitCode=143 Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.389722 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca3d78d6-072d-4a79-9325-aa054728bca4","Type":"ContainerDied","Data":"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136"} Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.389954 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca3d78d6-072d-4a79-9325-aa054728bca4","Type":"ContainerDied","Data":"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341"} Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.390065 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ca3d78d6-072d-4a79-9325-aa054728bca4","Type":"ContainerDied","Data":"274c0912c6d8ff2471472f68dc59458c71fa4a9560cbd2232878e010fa933f63"} Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.389675 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.422617 4492 scope.go:117] "RemoveContainer" containerID="83d595e0abde2bdad7d59fe4d7f42eb9afbe231390e9f8d9011d47b5da26af1a" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.443661 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.453234 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.470317 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.487551 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488070 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-api" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488092 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-api" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488107 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-notification-agent" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488116 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-notification-agent" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488145 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ccbd7a5-60ae-4701-8365-bf880455f53e" containerName="nova-manage" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488152 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ccbd7a5-60ae-4701-8365-bf880455f53e" containerName="nova-manage" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488161 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-central-agent" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488168 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-central-agent" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488176 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerName="init" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488182 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerName="init" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488194 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="sg-core" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488203 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="sg-core" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488217 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerName="dnsmasq-dns" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488223 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerName="dnsmasq-dns" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488236 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="proxy-httpd" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488242 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="proxy-httpd" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.488261 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-log" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488266 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-log" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488573 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-api" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488586 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="proxy-httpd" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488598 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" containerName="nova-api-log" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488609 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-notification-agent" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488618 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="26c40a07-b59c-4796-a5cb-caf6d8482de2" containerName="dnsmasq-dns" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488629 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="ceilometer-central-agent" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488639 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" containerName="sg-core" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.488650 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ccbd7a5-60ae-4701-8365-bf880455f53e" containerName="nova-manage" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.490601 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.493270 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.501093 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.501595 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.505739 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.521914 4492 scope.go:117] "RemoveContainer" containerID="b4d029bcec4c6834ecdadeb94cb29ccb139438b281e1e38aacb85375f963379f" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.528334 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.530721 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.532765 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.533131 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.533327 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.535275 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.549662 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.564927 4492 scope.go:117] "RemoveContainer" containerID="0ba6213840b64a97b086278a8e58a0b5d909e64748c1fbb84985473e03d96fdc" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.569453 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca3d78d6-072d-4a79-9325-aa054728bca4" path="/var/lib/kubelet/pods/ca3d78d6-072d-4a79-9325-aa054728bca4/volumes" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.570080 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0d824b3-b6dd-45d7-9675-0042da9a3048" path="/var/lib/kubelet/pods/f0d824b3-b6dd-45d7-9675-0042da9a3048/volumes" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.570818 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.596192 4492 scope.go:117] "RemoveContainer" containerID="e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630039 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlp6c\" (UniqueName: \"kubernetes.io/projected/a0e1abd4-7d2e-493b-b545-180dce022a33-kube-api-access-zlp6c\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630092 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630129 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630150 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-config-data\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630170 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e1abd4-7d2e-493b-b545-180dce022a33-log-httpd\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630316 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e1abd4-7d2e-493b-b545-180dce022a33-run-httpd\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630336 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.630388 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-scripts\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.666152 4492 scope.go:117] "RemoveContainer" containerID="3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.681302 4492 scope.go:117] "RemoveContainer" containerID="e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.681761 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136\": container with ID starting with e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136 not found: ID does not exist" containerID="e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.681845 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136"} err="failed to get container status \"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136\": rpc error: code = NotFound desc = could not find container \"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136\": container with ID starting with e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136 not found: ID does not exist" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.681879 4492 scope.go:117] "RemoveContainer" containerID="3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341" Feb 20 06:59:55 crc kubenswrapper[4492]: E0220 06:59:55.682184 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341\": container with ID starting with 3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341 not found: ID does not exist" containerID="3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.682219 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341"} err="failed to get container status \"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341\": rpc error: code = NotFound desc = could not find container \"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341\": container with ID starting with 3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341 not found: ID does not exist" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.682246 4492 scope.go:117] "RemoveContainer" containerID="e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.682652 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136"} err="failed to get container status \"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136\": rpc error: code = NotFound desc = could not find container \"e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136\": container with ID starting with e8a3d47d317bf632350641701de4fc427cc1f2cbad8161ad1a27ebc252bf3136 not found: ID does not exist" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.682685 4492 scope.go:117] "RemoveContainer" containerID="3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.682981 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341"} err="failed to get container status \"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341\": rpc error: code = NotFound desc = could not find container \"3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341\": container with ID starting with 3072987c8325c286754422ff469e7298f6d94ab1b83838fe39009c84d0c5d341 not found: ID does not exist" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.733799 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-config-data\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.733879 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.733903 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e1abd4-7d2e-493b-b545-180dce022a33-log-httpd\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.733926 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.734808 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e1abd4-7d2e-493b-b545-180dce022a33-run-httpd\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.734870 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.734905 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-public-tls-certs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.734965 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-config-data\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735008 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9591706-ed64-406e-b33f-d9f59db3c9c7-logs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735043 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-scripts\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735096 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlp6c\" (UniqueName: \"kubernetes.io/projected/a0e1abd4-7d2e-493b-b545-180dce022a33-kube-api-access-zlp6c\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735123 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsbkc\" (UniqueName: \"kubernetes.io/projected/e9591706-ed64-406e-b33f-d9f59db3c9c7-kube-api-access-jsbkc\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735147 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735191 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735319 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e1abd4-7d2e-493b-b545-180dce022a33-log-httpd\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.735577 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0e1abd4-7d2e-493b-b545-180dce022a33-run-httpd\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.739981 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-config-data\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.744555 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.759273 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.759451 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlp6c\" (UniqueName: \"kubernetes.io/projected/a0e1abd4-7d2e-493b-b545-180dce022a33-kube-api-access-zlp6c\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.759854 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.782520 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0e1abd4-7d2e-493b-b545-180dce022a33-scripts\") pod \"ceilometer-0\" (UID: \"a0e1abd4-7d2e-493b-b545-180dce022a33\") " pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.820804 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.837389 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-public-tls-certs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.837433 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-config-data\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.837468 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9591706-ed64-406e-b33f-d9f59db3c9c7-logs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.837548 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsbkc\" (UniqueName: \"kubernetes.io/projected/e9591706-ed64-406e-b33f-d9f59db3c9c7-kube-api-access-jsbkc\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.837591 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.837617 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.838561 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9591706-ed64-406e-b33f-d9f59db3c9c7-logs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.846965 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-public-tls-certs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.850024 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.850323 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-config-data\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.850368 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9591706-ed64-406e-b33f-d9f59db3c9c7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.855950 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsbkc\" (UniqueName: \"kubernetes.io/projected/e9591706-ed64-406e-b33f-d9f59db3c9c7-kube-api-access-jsbkc\") pod \"nova-api-0\" (UID: \"e9591706-ed64-406e-b33f-d9f59db3c9c7\") " pod="openstack/nova-api-0" Feb 20 06:59:55 crc kubenswrapper[4492]: I0220 06:59:55.861961 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 20 06:59:56 crc kubenswrapper[4492]: I0220 06:59:56.351194 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 20 06:59:56 crc kubenswrapper[4492]: I0220 06:59:56.407716 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e9591706-ed64-406e-b33f-d9f59db3c9c7","Type":"ContainerStarted","Data":"e01bddf0a35cd6aafe9f14d382466f6f2e09b7074bbb7093a80a6562e390da83"} Feb 20 06:59:56 crc kubenswrapper[4492]: W0220 06:59:56.410356 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0e1abd4_7d2e_493b_b545_180dce022a33.slice/crio-00efef4392663383da754a75d3f0dd0f2a1f3aaae75cb7d5dcb2d3f8ad1be352 WatchSource:0}: Error finding container 00efef4392663383da754a75d3f0dd0f2a1f3aaae75cb7d5dcb2d3f8ad1be352: Status 404 returned error can't find the container with id 00efef4392663383da754a75d3f0dd0f2a1f3aaae75cb7d5dcb2d3f8ad1be352 Feb 20 06:59:56 crc kubenswrapper[4492]: I0220 06:59:56.411495 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.007245 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.185680 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-combined-ca-bundle\") pod \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.186061 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-config-data\") pod \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.186214 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjht9\" (UniqueName: \"kubernetes.io/projected/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-kube-api-access-jjht9\") pod \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\" (UID: \"62fd389b-b2b1-4f40-b3cc-85be0c5c4650\") " Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.196527 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-kube-api-access-jjht9" (OuterVolumeSpecName: "kube-api-access-jjht9") pod "62fd389b-b2b1-4f40-b3cc-85be0c5c4650" (UID: "62fd389b-b2b1-4f40-b3cc-85be0c5c4650"). InnerVolumeSpecName "kube-api-access-jjht9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.207733 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62fd389b-b2b1-4f40-b3cc-85be0c5c4650" (UID: "62fd389b-b2b1-4f40-b3cc-85be0c5c4650"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.208132 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-config-data" (OuterVolumeSpecName: "config-data") pod "62fd389b-b2b1-4f40-b3cc-85be0c5c4650" (UID: "62fd389b-b2b1-4f40-b3cc-85be0c5c4650"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.298455 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.298509 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.298530 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjht9\" (UniqueName: \"kubernetes.io/projected/62fd389b-b2b1-4f40-b3cc-85be0c5c4650-kube-api-access-jjht9\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.434638 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e9591706-ed64-406e-b33f-d9f59db3c9c7","Type":"ContainerStarted","Data":"d5ceb6d1c7d511f17db80d4c48a384663499955ded19d4df07fcce7fb7368485"} Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.434689 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e9591706-ed64-406e-b33f-d9f59db3c9c7","Type":"ContainerStarted","Data":"30097d82445972e06b5fa2c4b28d76068e480ddefc4ff3f63df92beaff469139"} Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.442987 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e1abd4-7d2e-493b-b545-180dce022a33","Type":"ContainerStarted","Data":"ca6bed72287a51a01cab45c028b9bd38d01dbbd90301d8777be4b76e7e542f61"} Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.443031 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e1abd4-7d2e-493b-b545-180dce022a33","Type":"ContainerStarted","Data":"00efef4392663383da754a75d3f0dd0f2a1f3aaae75cb7d5dcb2d3f8ad1be352"} Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.447189 4492 generic.go:334] "Generic (PLEG): container finished" podID="62fd389b-b2b1-4f40-b3cc-85be0c5c4650" containerID="6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2" exitCode=0 Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.447229 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"62fd389b-b2b1-4f40-b3cc-85be0c5c4650","Type":"ContainerDied","Data":"6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2"} Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.447259 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"62fd389b-b2b1-4f40-b3cc-85be0c5c4650","Type":"ContainerDied","Data":"8909585513ea7feed0a0e725ce71c5cccf61017ac5e3aae3a4a47fc3e42ad3e2"} Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.447277 4492 scope.go:117] "RemoveContainer" containerID="6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.447270 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.470655 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.470627344 podStartE2EDuration="2.470627344s" podCreationTimestamp="2026-02-20 06:59:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:57.454149312 +0000 UTC m=+1154.225438290" watchObservedRunningTime="2026-02-20 06:59:57.470627344 +0000 UTC m=+1154.241916341" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.493459 4492 scope.go:117] "RemoveContainer" containerID="6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2" Feb 20 06:59:57 crc kubenswrapper[4492]: E0220 06:59:57.500311 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2\": container with ID starting with 6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2 not found: ID does not exist" containerID="6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.500357 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2"} err="failed to get container status \"6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2\": rpc error: code = NotFound desc = could not find container \"6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2\": container with ID starting with 6898422b4fea07a1b1111194b17bcf1047951006a94f6bcc958fad3ff82cd9a2 not found: ID does not exist" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.504092 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.518900 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.528686 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:57 crc kubenswrapper[4492]: E0220 06:59:57.529069 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fd389b-b2b1-4f40-b3cc-85be0c5c4650" containerName="nova-scheduler-scheduler" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.529091 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fd389b-b2b1-4f40-b3cc-85be0c5c4650" containerName="nova-scheduler-scheduler" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.529280 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fd389b-b2b1-4f40-b3cc-85be0c5c4650" containerName="nova-scheduler-scheduler" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.530011 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.534000 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.539009 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.565580 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62fd389b-b2b1-4f40-b3cc-85be0c5c4650" path="/var/lib/kubelet/pods/62fd389b-b2b1-4f40-b3cc-85be0c5c4650/volumes" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.604029 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.604070 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-config-data\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.604131 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ljx5\" (UniqueName: \"kubernetes.io/projected/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-kube-api-access-4ljx5\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.705136 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.705373 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-config-data\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.705490 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ljx5\" (UniqueName: \"kubernetes.io/projected/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-kube-api-access-4ljx5\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.711284 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-config-data\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.711356 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.734881 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ljx5\" (UniqueName: \"kubernetes.io/projected/1c8d2876-8ebc-46af-aa6a-5b03bb8ee207-kube-api-access-4ljx5\") pod \"nova-scheduler-0\" (UID: \"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207\") " pod="openstack/nova-scheduler-0" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.738115 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": read tcp 10.217.0.2:47426->10.217.0.210:8775: read: connection reset by peer" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.738141 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.210:8775/\": read tcp 10.217.0.2:47430->10.217.0.210:8775: read: connection reset by peer" Feb 20 06:59:57 crc kubenswrapper[4492]: I0220 06:59:57.850947 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.104915 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.125126 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-nova-metadata-tls-certs\") pod \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.125168 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle\") pod \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.125436 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9z9g\" (UniqueName: \"kubernetes.io/projected/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-kube-api-access-v9z9g\") pod \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.125488 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-logs\") pod \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.125734 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-config-data\") pod \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.139766 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-logs" (OuterVolumeSpecName: "logs") pod "2127cdef-bc2d-46fe-8f3e-0934df26b4d6" (UID: "2127cdef-bc2d-46fe-8f3e-0934df26b4d6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.199523 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-kube-api-access-v9z9g" (OuterVolumeSpecName: "kube-api-access-v9z9g") pod "2127cdef-bc2d-46fe-8f3e-0934df26b4d6" (UID: "2127cdef-bc2d-46fe-8f3e-0934df26b4d6"). InnerVolumeSpecName "kube-api-access-v9z9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.216706 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-config-data" (OuterVolumeSpecName: "config-data") pod "2127cdef-bc2d-46fe-8f3e-0934df26b4d6" (UID: "2127cdef-bc2d-46fe-8f3e-0934df26b4d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.227237 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2127cdef-bc2d-46fe-8f3e-0934df26b4d6" (UID: "2127cdef-bc2d-46fe-8f3e-0934df26b4d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.228490 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle\") pod \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\" (UID: \"2127cdef-bc2d-46fe-8f3e-0934df26b4d6\") " Feb 20 06:59:58 crc kubenswrapper[4492]: W0220 06:59:58.229360 4492 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/2127cdef-bc2d-46fe-8f3e-0934df26b4d6/volumes/kubernetes.io~secret/combined-ca-bundle Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.229400 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2127cdef-bc2d-46fe-8f3e-0934df26b4d6" (UID: "2127cdef-bc2d-46fe-8f3e-0934df26b4d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.234649 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.234676 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9z9g\" (UniqueName: \"kubernetes.io/projected/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-kube-api-access-v9z9g\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.234689 4492 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-logs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.234716 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.255554 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2127cdef-bc2d-46fe-8f3e-0934df26b4d6" (UID: "2127cdef-bc2d-46fe-8f3e-0934df26b4d6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.287756 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.338055 4492 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2127cdef-bc2d-46fe-8f3e-0934df26b4d6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.466001 4492 generic.go:334] "Generic (PLEG): container finished" podID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerID="d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2" exitCode=0 Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.466128 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.466157 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2127cdef-bc2d-46fe-8f3e-0934df26b4d6","Type":"ContainerDied","Data":"d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2"} Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.466753 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2127cdef-bc2d-46fe-8f3e-0934df26b4d6","Type":"ContainerDied","Data":"6d85aebed9ac0a8b4ed28aa7e4fb96d90a049b52588342e1e3ee87c7f2b80b45"} Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.466803 4492 scope.go:117] "RemoveContainer" containerID="d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.472214 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207","Type":"ContainerStarted","Data":"d2dfa59015da57b7e6da7a204cde0141ac8a374b9159ed5c1acafd16191ed1a4"} Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.472257 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1c8d2876-8ebc-46af-aa6a-5b03bb8ee207","Type":"ContainerStarted","Data":"2c15778b388a49b99bb13c3b980dbef994e2877ad96df3dbf0e8443717c2e5a6"} Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.477525 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e1abd4-7d2e-493b-b545-180dce022a33","Type":"ContainerStarted","Data":"9efcc1ca32f8c03c796297e69e628ca3781544703f0b340f9857741df6f0577b"} Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.492810 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.4927900379999999 podStartE2EDuration="1.492790038s" podCreationTimestamp="2026-02-20 06:59:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 06:59:58.487582603 +0000 UTC m=+1155.258871580" watchObservedRunningTime="2026-02-20 06:59:58.492790038 +0000 UTC m=+1155.264079016" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.513174 4492 scope.go:117] "RemoveContainer" containerID="26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.526514 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.543284 4492 scope.go:117] "RemoveContainer" containerID="d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2" Feb 20 06:59:58 crc kubenswrapper[4492]: E0220 06:59:58.546071 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2\": container with ID starting with d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2 not found: ID does not exist" containerID="d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.546119 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2"} err="failed to get container status \"d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2\": rpc error: code = NotFound desc = could not find container \"d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2\": container with ID starting with d1ef448c5ef202c1e1c360812f973e6d18691cd3e1292e33ee8f6984797da2f2 not found: ID does not exist" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.546154 4492 scope.go:117] "RemoveContainer" containerID="26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.552527 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:58 crc kubenswrapper[4492]: E0220 06:59:58.552725 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c\": container with ID starting with 26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c not found: ID does not exist" containerID="26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.552819 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c"} err="failed to get container status \"26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c\": rpc error: code = NotFound desc = could not find container \"26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c\": container with ID starting with 26efa60be6b0bba1e4021694fffa022afbe24791629f6312543c9f389a67f07c not found: ID does not exist" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.568365 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:58 crc kubenswrapper[4492]: E0220 06:59:58.569235 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-metadata" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.569302 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-metadata" Feb 20 06:59:58 crc kubenswrapper[4492]: E0220 06:59:58.569358 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-log" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.569400 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-log" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.569848 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-log" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.569964 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" containerName="nova-metadata-metadata" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.571600 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.576344 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.590463 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.590745 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.646398 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkpfg\" (UniqueName: \"kubernetes.io/projected/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-kube-api-access-lkpfg\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.646540 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.646682 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.646825 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-config-data\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.646943 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-logs\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.748898 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.749059 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-config-data\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.749189 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-logs\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.749712 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-logs\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.749848 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkpfg\" (UniqueName: \"kubernetes.io/projected/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-kube-api-access-lkpfg\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.749882 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.753338 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-config-data\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.753356 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.755171 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.782495 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkpfg\" (UniqueName: \"kubernetes.io/projected/47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a-kube-api-access-lkpfg\") pod \"nova-metadata-0\" (UID: \"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a\") " pod="openstack/nova-metadata-0" Feb 20 06:59:58 crc kubenswrapper[4492]: I0220 06:59:58.968673 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 20 06:59:59 crc kubenswrapper[4492]: I0220 06:59:59.408590 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 20 06:59:59 crc kubenswrapper[4492]: W0220 06:59:59.418351 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47f7dfdc_8c6b_4dcb_83e0_0a1ee209f03a.slice/crio-ee0500c4ef79e34cdeaba020f226abee81891517dcbc5de071127ae32dfae9d1 WatchSource:0}: Error finding container ee0500c4ef79e34cdeaba020f226abee81891517dcbc5de071127ae32dfae9d1: Status 404 returned error can't find the container with id ee0500c4ef79e34cdeaba020f226abee81891517dcbc5de071127ae32dfae9d1 Feb 20 06:59:59 crc kubenswrapper[4492]: I0220 06:59:59.490193 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a","Type":"ContainerStarted","Data":"ee0500c4ef79e34cdeaba020f226abee81891517dcbc5de071127ae32dfae9d1"} Feb 20 06:59:59 crc kubenswrapper[4492]: I0220 06:59:59.496798 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e1abd4-7d2e-493b-b545-180dce022a33","Type":"ContainerStarted","Data":"8fae96a7a065d49903d5e911e9d9d2fa5717fc4bb408cc506594bb965fe687ea"} Feb 20 06:59:59 crc kubenswrapper[4492]: I0220 06:59:59.583044 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2127cdef-bc2d-46fe-8f3e-0934df26b4d6" path="/var/lib/kubelet/pods/2127cdef-bc2d-46fe-8f3e-0934df26b4d6/volumes" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.161615 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms"] Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.163042 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.165675 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.170952 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.171901 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms"] Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.189025 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmg2h\" (UniqueName: \"kubernetes.io/projected/09ad0831-93b1-4f58-9a4b-772d16f072c2-kube-api-access-vmg2h\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.189365 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09ad0831-93b1-4f58-9a4b-772d16f072c2-config-volume\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.189432 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09ad0831-93b1-4f58-9a4b-772d16f072c2-secret-volume\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.291799 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09ad0831-93b1-4f58-9a4b-772d16f072c2-config-volume\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.291946 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09ad0831-93b1-4f58-9a4b-772d16f072c2-secret-volume\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.292167 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmg2h\" (UniqueName: \"kubernetes.io/projected/09ad0831-93b1-4f58-9a4b-772d16f072c2-kube-api-access-vmg2h\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.292729 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09ad0831-93b1-4f58-9a4b-772d16f072c2-config-volume\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.295619 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09ad0831-93b1-4f58-9a4b-772d16f072c2-secret-volume\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.308437 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmg2h\" (UniqueName: \"kubernetes.io/projected/09ad0831-93b1-4f58-9a4b-772d16f072c2-kube-api-access-vmg2h\") pod \"collect-profiles-29526180-lqsms\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.509857 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0e1abd4-7d2e-493b-b545-180dce022a33","Type":"ContainerStarted","Data":"ec52ae0119173b98a4f8ad03d6a3c30649153bdc631bdfd21352d97ecfca2964"} Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.510351 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.513089 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a","Type":"ContainerStarted","Data":"a129ca131aefee3644a89d080578e9eeed481dd30a5ee9941c8313aa3a113651"} Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.513148 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a","Type":"ContainerStarted","Data":"9c59ecb67fb97015ae28a3f03fb9961e3af2ad136f11256a1beefdc31327a013"} Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.534571 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.858683753 podStartE2EDuration="5.534551343s" podCreationTimestamp="2026-02-20 06:59:55 +0000 UTC" firstStartedPulling="2026-02-20 06:59:56.418686136 +0000 UTC m=+1153.189975114" lastFinishedPulling="2026-02-20 07:00:00.094553726 +0000 UTC m=+1156.865842704" observedRunningTime="2026-02-20 07:00:00.529192281 +0000 UTC m=+1157.300481259" watchObservedRunningTime="2026-02-20 07:00:00.534551343 +0000 UTC m=+1157.305840321" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.567562 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.567533476 podStartE2EDuration="2.567533476s" podCreationTimestamp="2026-02-20 06:59:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:00:00.561702214 +0000 UTC m=+1157.332991191" watchObservedRunningTime="2026-02-20 07:00:00.567533476 +0000 UTC m=+1157.338822454" Feb 20 07:00:00 crc kubenswrapper[4492]: I0220 07:00:00.582529 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:01 crc kubenswrapper[4492]: W0220 07:00:01.062217 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09ad0831_93b1_4f58_9a4b_772d16f072c2.slice/crio-98a16f8adfad220dfb891bc0adb89faab4525125110c2f8d50cdf8306ab6785c WatchSource:0}: Error finding container 98a16f8adfad220dfb891bc0adb89faab4525125110c2f8d50cdf8306ab6785c: Status 404 returned error can't find the container with id 98a16f8adfad220dfb891bc0adb89faab4525125110c2f8d50cdf8306ab6785c Feb 20 07:00:01 crc kubenswrapper[4492]: I0220 07:00:01.070413 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms"] Feb 20 07:00:01 crc kubenswrapper[4492]: I0220 07:00:01.527555 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" event={"ID":"09ad0831-93b1-4f58-9a4b-772d16f072c2","Type":"ContainerStarted","Data":"a82787a5e86af1abdaa5700884c2a68da34c7a9702d857e4cc1571e00a22be5b"} Feb 20 07:00:01 crc kubenswrapper[4492]: I0220 07:00:01.527854 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" event={"ID":"09ad0831-93b1-4f58-9a4b-772d16f072c2","Type":"ContainerStarted","Data":"98a16f8adfad220dfb891bc0adb89faab4525125110c2f8d50cdf8306ab6785c"} Feb 20 07:00:01 crc kubenswrapper[4492]: I0220 07:00:01.547900 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" podStartSLOduration=1.5478840630000001 podStartE2EDuration="1.547884063s" podCreationTimestamp="2026-02-20 07:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:00:01.541077783 +0000 UTC m=+1158.312366761" watchObservedRunningTime="2026-02-20 07:00:01.547884063 +0000 UTC m=+1158.319173041" Feb 20 07:00:02 crc kubenswrapper[4492]: I0220 07:00:02.541568 4492 generic.go:334] "Generic (PLEG): container finished" podID="09ad0831-93b1-4f58-9a4b-772d16f072c2" containerID="a82787a5e86af1abdaa5700884c2a68da34c7a9702d857e4cc1571e00a22be5b" exitCode=0 Feb 20 07:00:02 crc kubenswrapper[4492]: I0220 07:00:02.541682 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" event={"ID":"09ad0831-93b1-4f58-9a4b-772d16f072c2","Type":"ContainerDied","Data":"a82787a5e86af1abdaa5700884c2a68da34c7a9702d857e4cc1571e00a22be5b"} Feb 20 07:00:02 crc kubenswrapper[4492]: I0220 07:00:02.851397 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 20 07:00:03 crc kubenswrapper[4492]: I0220 07:00:03.897243 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:03 crc kubenswrapper[4492]: I0220 07:00:03.969696 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 07:00:03 crc kubenswrapper[4492]: I0220 07:00:03.969756 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.002835 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09ad0831-93b1-4f58-9a4b-772d16f072c2-secret-volume\") pod \"09ad0831-93b1-4f58-9a4b-772d16f072c2\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.003082 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09ad0831-93b1-4f58-9a4b-772d16f072c2-config-volume\") pod \"09ad0831-93b1-4f58-9a4b-772d16f072c2\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.003147 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmg2h\" (UniqueName: \"kubernetes.io/projected/09ad0831-93b1-4f58-9a4b-772d16f072c2-kube-api-access-vmg2h\") pod \"09ad0831-93b1-4f58-9a4b-772d16f072c2\" (UID: \"09ad0831-93b1-4f58-9a4b-772d16f072c2\") " Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.004074 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ad0831-93b1-4f58-9a4b-772d16f072c2-config-volume" (OuterVolumeSpecName: "config-volume") pod "09ad0831-93b1-4f58-9a4b-772d16f072c2" (UID: "09ad0831-93b1-4f58-9a4b-772d16f072c2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.005825 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09ad0831-93b1-4f58-9a4b-772d16f072c2-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.021273 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ad0831-93b1-4f58-9a4b-772d16f072c2-kube-api-access-vmg2h" (OuterVolumeSpecName: "kube-api-access-vmg2h") pod "09ad0831-93b1-4f58-9a4b-772d16f072c2" (UID: "09ad0831-93b1-4f58-9a4b-772d16f072c2"). InnerVolumeSpecName "kube-api-access-vmg2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.021398 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ad0831-93b1-4f58-9a4b-772d16f072c2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09ad0831-93b1-4f58-9a4b-772d16f072c2" (UID: "09ad0831-93b1-4f58-9a4b-772d16f072c2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.108611 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmg2h\" (UniqueName: \"kubernetes.io/projected/09ad0831-93b1-4f58-9a4b-772d16f072c2-kube-api-access-vmg2h\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.108938 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09ad0831-93b1-4f58-9a4b-772d16f072c2-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.579509 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" event={"ID":"09ad0831-93b1-4f58-9a4b-772d16f072c2","Type":"ContainerDied","Data":"98a16f8adfad220dfb891bc0adb89faab4525125110c2f8d50cdf8306ab6785c"} Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.579566 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98a16f8adfad220dfb891bc0adb89faab4525125110c2f8d50cdf8306ab6785c" Feb 20 07:00:04 crc kubenswrapper[4492]: I0220 07:00:04.579636 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms" Feb 20 07:00:05 crc kubenswrapper[4492]: I0220 07:00:05.863521 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 07:00:05 crc kubenswrapper[4492]: I0220 07:00:05.863863 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 20 07:00:06 crc kubenswrapper[4492]: I0220 07:00:06.877631 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e9591706-ed64-406e-b33f-d9f59db3c9c7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 07:00:06 crc kubenswrapper[4492]: I0220 07:00:06.878283 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e9591706-ed64-406e-b33f-d9f59db3c9c7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 07:00:07 crc kubenswrapper[4492]: I0220 07:00:07.852215 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 20 07:00:07 crc kubenswrapper[4492]: I0220 07:00:07.874665 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 20 07:00:08 crc kubenswrapper[4492]: I0220 07:00:08.659772 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 20 07:00:08 crc kubenswrapper[4492]: I0220 07:00:08.970077 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 07:00:08 crc kubenswrapper[4492]: I0220 07:00:08.970134 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.311580 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.311663 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.311720 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.312447 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1968b24edc23a5de4ecabc4dcc18c03f061c8965af1f7e77792ae0d80771d5c6"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.312531 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://1968b24edc23a5de4ecabc4dcc18c03f061c8965af1f7e77792ae0d80771d5c6" gracePeriod=600 Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.655614 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="1968b24edc23a5de4ecabc4dcc18c03f061c8965af1f7e77792ae0d80771d5c6" exitCode=0 Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.655738 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"1968b24edc23a5de4ecabc4dcc18c03f061c8965af1f7e77792ae0d80771d5c6"} Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.655857 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"ab8284c72f81bddf931f4abee5f04c18596ed9e487d4c6a5d3e615ec918dcbba"} Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.655920 4492 scope.go:117] "RemoveContainer" containerID="721ada477bd6e659aca9b99823e1dcf15d8562664d9f214bcd56e2ac3871cc6b" Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.985900 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.222:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 07:00:09 crc kubenswrapper[4492]: I0220 07:00:09.985911 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.222:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 20 07:00:15 crc kubenswrapper[4492]: I0220 07:00:15.870543 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 07:00:15 crc kubenswrapper[4492]: I0220 07:00:15.871263 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 20 07:00:15 crc kubenswrapper[4492]: I0220 07:00:15.871852 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 07:00:15 crc kubenswrapper[4492]: I0220 07:00:15.873126 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 20 07:00:15 crc kubenswrapper[4492]: I0220 07:00:15.880838 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 07:00:15 crc kubenswrapper[4492]: I0220 07:00:15.882792 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 20 07:00:18 crc kubenswrapper[4492]: I0220 07:00:18.978836 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 07:00:18 crc kubenswrapper[4492]: I0220 07:00:18.979551 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 20 07:00:18 crc kubenswrapper[4492]: I0220 07:00:18.985434 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 07:00:18 crc kubenswrapper[4492]: I0220 07:00:18.986594 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 20 07:00:25 crc kubenswrapper[4492]: I0220 07:00:25.834100 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 20 07:00:34 crc kubenswrapper[4492]: I0220 07:00:34.145072 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 07:00:35 crc kubenswrapper[4492]: I0220 07:00:35.102345 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 07:00:39 crc kubenswrapper[4492]: I0220 07:00:39.044640 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerName="rabbitmq" containerID="cri-o://3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b" gracePeriod=604796 Feb 20 07:00:39 crc kubenswrapper[4492]: I0220 07:00:39.895394 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerName="rabbitmq" containerID="cri-o://9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf" gracePeriod=604796 Feb 20 07:00:44 crc kubenswrapper[4492]: I0220 07:00:44.503193 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.94:5671: connect: connection refused" Feb 20 07:00:44 crc kubenswrapper[4492]: I0220 07:00:44.597899 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.95:5671: connect: connection refused" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.200818 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57486d48df-tv7hf"] Feb 20 07:00:45 crc kubenswrapper[4492]: E0220 07:00:45.201717 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09ad0831-93b1-4f58-9a4b-772d16f072c2" containerName="collect-profiles" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.201802 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="09ad0831-93b1-4f58-9a4b-772d16f072c2" containerName="collect-profiles" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.202102 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="09ad0831-93b1-4f58-9a4b-772d16f072c2" containerName="collect-profiles" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.203230 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.210156 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.232332 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57486d48df-tv7hf"] Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.243948 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2dfm\" (UniqueName: \"kubernetes.io/projected/f7408165-ed60-4ff1-a838-a5e98cd2710f-kube-api-access-q2dfm\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.243994 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-config\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.244026 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-nb\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.244096 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-sb\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.244117 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-svc\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.244136 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-openstack-edpm-ipam\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.244191 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-swift-storage-0\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.347237 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2dfm\" (UniqueName: \"kubernetes.io/projected/f7408165-ed60-4ff1-a838-a5e98cd2710f-kube-api-access-q2dfm\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.347315 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-config\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.347388 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-nb\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.347644 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-sb\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.347684 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-svc\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.347714 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-openstack-edpm-ipam\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.347908 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-swift-storage-0\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.348629 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-nb\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.349414 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-swift-storage-0\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.349538 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-openstack-edpm-ipam\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.350063 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-sb\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.350093 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-config\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.351000 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-svc\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.372777 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2dfm\" (UniqueName: \"kubernetes.io/projected/f7408165-ed60-4ff1-a838-a5e98cd2710f-kube-api-access-q2dfm\") pod \"dnsmasq-dns-57486d48df-tv7hf\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.530909 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.682505 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.756550 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2719d90-a43e-4b5a-b4ad-e948f75c763a-pod-info\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.756650 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mvg6\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-kube-api-access-4mvg6\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758338 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2719d90-a43e-4b5a-b4ad-e948f75c763a-erlang-cookie-secret\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758498 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-config-data\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758560 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-server-conf\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758583 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-erlang-cookie\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758656 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-plugins-conf\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758682 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-confd\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758729 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-tls\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758757 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.758822 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-plugins\") pod \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\" (UID: \"a2719d90-a43e-4b5a-b4ad-e948f75c763a\") " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.766317 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.772280 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.788528 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.790057 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a2719d90-a43e-4b5a-b4ad-e948f75c763a-pod-info" (OuterVolumeSpecName: "pod-info") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.791687 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.793439 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2719d90-a43e-4b5a-b4ad-e948f75c763a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.793538 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.794253 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-kube-api-access-4mvg6" (OuterVolumeSpecName: "kube-api-access-4mvg6") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "kube-api-access-4mvg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.819976 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-config-data" (OuterVolumeSpecName: "config-data") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867072 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867400 4492 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867459 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867568 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867662 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867746 4492 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2719d90-a43e-4b5a-b4ad-e948f75c763a-pod-info\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867825 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mvg6\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-kube-api-access-4mvg6\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867897 4492 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2719d90-a43e-4b5a-b4ad-e948f75c763a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.867976 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.884204 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-server-conf" (OuterVolumeSpecName: "server-conf") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.887626 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.932091 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a2719d90-a43e-4b5a-b4ad-e948f75c763a" (UID: "a2719d90-a43e-4b5a-b4ad-e948f75c763a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.974021 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2719d90-a43e-4b5a-b4ad-e948f75c763a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.974050 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:45 crc kubenswrapper[4492]: I0220 07:00:45.974061 4492 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2719d90-a43e-4b5a-b4ad-e948f75c763a-server-conf\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.013754 4492 generic.go:334] "Generic (PLEG): container finished" podID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerID="3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b" exitCode=0 Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.013807 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a2719d90-a43e-4b5a-b4ad-e948f75c763a","Type":"ContainerDied","Data":"3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b"} Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.013844 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a2719d90-a43e-4b5a-b4ad-e948f75c763a","Type":"ContainerDied","Data":"aa76559b20f9369723626daba80900b006c6c6528401ac3da0d02869127ce79c"} Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.013863 4492 scope.go:117] "RemoveContainer" containerID="3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.014010 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.070010 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.072559 4492 scope.go:117] "RemoveContainer" containerID="89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.083597 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.089634 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 07:00:46 crc kubenswrapper[4492]: E0220 07:00:46.090738 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerName="rabbitmq" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.090933 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerName="rabbitmq" Feb 20 07:00:46 crc kubenswrapper[4492]: E0220 07:00:46.092223 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerName="setup-container" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.092280 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerName="setup-container" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.092607 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" containerName="rabbitmq" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.093759 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.104637 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.104918 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.105145 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tnqs6" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.105295 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.105543 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.105738 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.105163 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.148549 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.154542 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57486d48df-tv7hf"] Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.164972 4492 scope.go:117] "RemoveContainer" containerID="3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b" Feb 20 07:00:46 crc kubenswrapper[4492]: E0220 07:00:46.165941 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b\": container with ID starting with 3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b not found: ID does not exist" containerID="3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.166024 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b"} err="failed to get container status \"3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b\": rpc error: code = NotFound desc = could not find container \"3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b\": container with ID starting with 3075cb3db23fdc847f5755644376972f996a10673cf0dafdd5465ff6f4d3610b not found: ID does not exist" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.166091 4492 scope.go:117] "RemoveContainer" containerID="89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a" Feb 20 07:00:46 crc kubenswrapper[4492]: E0220 07:00:46.176694 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a\": container with ID starting with 89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a not found: ID does not exist" containerID="89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.176739 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a"} err="failed to get container status \"89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a\": rpc error: code = NotFound desc = could not find container \"89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a\": container with ID starting with 89f3aac15a32d0558b3d7a146b33041eac27296401d9a99787873dad298d0a4a not found: ID does not exist" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.182967 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dfa138-8cc8-40e0-b2df-21449a994146-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183113 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183133 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183152 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183189 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-config-data\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183229 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183445 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183515 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183538 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n47jr\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-kube-api-access-n47jr\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183598 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dfa138-8cc8-40e0-b2df-21449a994146-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.183635 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: W0220 07:00:46.187822 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7408165_ed60_4ff1_a838_a5e98cd2710f.slice/crio-3ca7f5a72ad5764b76696bce9a6df90793eed5f5364af549548f61279ef10daa WatchSource:0}: Error finding container 3ca7f5a72ad5764b76696bce9a6df90793eed5f5364af549548f61279ef10daa: Status 404 returned error can't find the container with id 3ca7f5a72ad5764b76696bce9a6df90793eed5f5364af549548f61279ef10daa Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288063 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288119 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n47jr\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-kube-api-access-n47jr\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288182 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dfa138-8cc8-40e0-b2df-21449a994146-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288228 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288269 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dfa138-8cc8-40e0-b2df-21449a994146-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288422 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288466 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288516 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288568 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-config-data\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288602 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.288654 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.289712 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.291105 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.293066 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a4dfa138-8cc8-40e0-b2df-21449a994146-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.295190 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.307926 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a4dfa138-8cc8-40e0-b2df-21449a994146-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.310014 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.310405 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-config-data\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.319406 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a4dfa138-8cc8-40e0-b2df-21449a994146-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.340177 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n47jr\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-kube-api-access-n47jr\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.341569 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.342324 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a4dfa138-8cc8-40e0-b2df-21449a994146-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.370067 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"a4dfa138-8cc8-40e0-b2df-21449a994146\") " pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.431747 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.530584 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.707435 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.707847 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-plugins-conf\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.707894 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c936d390-aaa6-4904-a1ca-357e2dac4cde-pod-info\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.707916 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c936d390-aaa6-4904-a1ca-357e2dac4cde-erlang-cookie-secret\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.708036 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-config-data\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.708058 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-server-conf\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.708124 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-tls\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.708162 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpjhf\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-kube-api-access-hpjhf\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.708200 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-erlang-cookie\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.708224 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-plugins\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.708271 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-confd\") pod \"c936d390-aaa6-4904-a1ca-357e2dac4cde\" (UID: \"c936d390-aaa6-4904-a1ca-357e2dac4cde\") " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.710888 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.712291 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.713782 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.717850 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/c936d390-aaa6-4904-a1ca-357e2dac4cde-pod-info" (OuterVolumeSpecName: "pod-info") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.725762 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-kube-api-access-hpjhf" (OuterVolumeSpecName: "kube-api-access-hpjhf") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "kube-api-access-hpjhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.725873 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.726881 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.736987 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c936d390-aaa6-4904-a1ca-357e2dac4cde-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.741709 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-config-data" (OuterVolumeSpecName: "config-data") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.781123 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-server-conf" (OuterVolumeSpecName: "server-conf") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811159 4492 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c936d390-aaa6-4904-a1ca-357e2dac4cde-pod-info\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811262 4492 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c936d390-aaa6-4904-a1ca-357e2dac4cde-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811560 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811626 4492 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-server-conf\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811693 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811740 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpjhf\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-kube-api-access-hpjhf\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811818 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811882 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.811962 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.812027 4492 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c936d390-aaa6-4904-a1ca-357e2dac4cde-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.828908 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.831113 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "c936d390-aaa6-4904-a1ca-357e2dac4cde" (UID: "c936d390-aaa6-4904-a1ca-357e2dac4cde"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.914262 4492 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c936d390-aaa6-4904-a1ca-357e2dac4cde-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.914298 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:46 crc kubenswrapper[4492]: W0220 07:00:46.933389 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4dfa138_8cc8_40e0_b2df_21449a994146.slice/crio-2f73c6864d94ed85ea64ae82edf1e5c0da5ed4e7bd5eff9fa4918649fd07d018 WatchSource:0}: Error finding container 2f73c6864d94ed85ea64ae82edf1e5c0da5ed4e7bd5eff9fa4918649fd07d018: Status 404 returned error can't find the container with id 2f73c6864d94ed85ea64ae82edf1e5c0da5ed4e7bd5eff9fa4918649fd07d018 Feb 20 07:00:46 crc kubenswrapper[4492]: I0220 07:00:46.934102 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.027696 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a4dfa138-8cc8-40e0-b2df-21449a994146","Type":"ContainerStarted","Data":"2f73c6864d94ed85ea64ae82edf1e5c0da5ed4e7bd5eff9fa4918649fd07d018"} Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.029612 4492 generic.go:334] "Generic (PLEG): container finished" podID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerID="75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a" exitCode=0 Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.029722 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" event={"ID":"f7408165-ed60-4ff1-a838-a5e98cd2710f","Type":"ContainerDied","Data":"75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a"} Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.029811 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" event={"ID":"f7408165-ed60-4ff1-a838-a5e98cd2710f","Type":"ContainerStarted","Data":"3ca7f5a72ad5764b76696bce9a6df90793eed5f5364af549548f61279ef10daa"} Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.032201 4492 generic.go:334] "Generic (PLEG): container finished" podID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerID="9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf" exitCode=0 Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.032258 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c936d390-aaa6-4904-a1ca-357e2dac4cde","Type":"ContainerDied","Data":"9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf"} Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.032292 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c936d390-aaa6-4904-a1ca-357e2dac4cde","Type":"ContainerDied","Data":"6f65f0257cb727f9b649ed38ca7978d001b557c6a531ae174240b3924719bf89"} Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.032318 4492 scope.go:117] "RemoveContainer" containerID="9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.032466 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.071364 4492 scope.go:117] "RemoveContainer" containerID="6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.101109 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.106208 4492 scope.go:117] "RemoveContainer" containerID="9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf" Feb 20 07:00:47 crc kubenswrapper[4492]: E0220 07:00:47.107458 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf\": container with ID starting with 9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf not found: ID does not exist" containerID="9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.107823 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf"} err="failed to get container status \"9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf\": rpc error: code = NotFound desc = could not find container \"9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf\": container with ID starting with 9ede01d0268eaf85294753df1e1c0d8909b3c49b611725ed7b316db0e056b9bf not found: ID does not exist" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.107943 4492 scope.go:117] "RemoveContainer" containerID="6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449" Feb 20 07:00:47 crc kubenswrapper[4492]: E0220 07:00:47.110532 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449\": container with ID starting with 6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449 not found: ID does not exist" containerID="6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.110629 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449"} err="failed to get container status \"6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449\": rpc error: code = NotFound desc = could not find container \"6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449\": container with ID starting with 6fd2329b9ac92f30f97c75e7533d4b94be08a53b7038087c55b9b7589d70d449 not found: ID does not exist" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.118964 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.127196 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 07:00:47 crc kubenswrapper[4492]: E0220 07:00:47.127765 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerName="setup-container" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.127786 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerName="setup-container" Feb 20 07:00:47 crc kubenswrapper[4492]: E0220 07:00:47.127812 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerName="rabbitmq" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.127819 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerName="rabbitmq" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.128070 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" containerName="rabbitmq" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.129209 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.132536 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.132703 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wtkz5" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.135646 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.135834 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.135915 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.136006 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.136081 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.136204 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.328788 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.328843 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18f922e8-15c1-49a7-af1b-38a85b924404-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.328939 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.328989 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.329103 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.329132 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.329197 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.329221 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7wbr\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-kube-api-access-n7wbr\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.329689 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18f922e8-15c1-49a7-af1b-38a85b924404-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.329744 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.329876 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432541 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432592 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432622 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432647 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7wbr\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-kube-api-access-n7wbr\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432718 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18f922e8-15c1-49a7-af1b-38a85b924404-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432736 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432763 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432807 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432831 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18f922e8-15c1-49a7-af1b-38a85b924404-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432855 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.432903 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.433338 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.433944 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.434206 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.435285 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.436283 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/18f922e8-15c1-49a7-af1b-38a85b924404-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.436391 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.440391 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/18f922e8-15c1-49a7-af1b-38a85b924404-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.441165 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.441261 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.455096 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7wbr\" (UniqueName: \"kubernetes.io/projected/18f922e8-15c1-49a7-af1b-38a85b924404-kube-api-access-n7wbr\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.457711 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/18f922e8-15c1-49a7-af1b-38a85b924404-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.467178 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"18f922e8-15c1-49a7-af1b-38a85b924404\") " pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.572020 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2719d90-a43e-4b5a-b4ad-e948f75c763a" path="/var/lib/kubelet/pods/a2719d90-a43e-4b5a-b4ad-e948f75c763a/volumes" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.573413 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c936d390-aaa6-4904-a1ca-357e2dac4cde" path="/var/lib/kubelet/pods/c936d390-aaa6-4904-a1ca-357e2dac4cde/volumes" Feb 20 07:00:47 crc kubenswrapper[4492]: I0220 07:00:47.749293 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:00:48 crc kubenswrapper[4492]: I0220 07:00:48.050158 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" event={"ID":"f7408165-ed60-4ff1-a838-a5e98cd2710f","Type":"ContainerStarted","Data":"1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491"} Feb 20 07:00:48 crc kubenswrapper[4492]: I0220 07:00:48.050466 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:48 crc kubenswrapper[4492]: I0220 07:00:48.076130 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" podStartSLOduration=3.076109666 podStartE2EDuration="3.076109666s" podCreationTimestamp="2026-02-20 07:00:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:00:48.066690455 +0000 UTC m=+1204.837979433" watchObservedRunningTime="2026-02-20 07:00:48.076109666 +0000 UTC m=+1204.847398634" Feb 20 07:00:48 crc kubenswrapper[4492]: I0220 07:00:48.154757 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 20 07:00:49 crc kubenswrapper[4492]: I0220 07:00:49.063787 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18f922e8-15c1-49a7-af1b-38a85b924404","Type":"ContainerStarted","Data":"a6b8442e82f44331d4b7011e3ec76c879f7e057de92c62d632ce0b45723f5fd7"} Feb 20 07:00:49 crc kubenswrapper[4492]: I0220 07:00:49.066201 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a4dfa138-8cc8-40e0-b2df-21449a994146","Type":"ContainerStarted","Data":"5cdc5ebab8d08f295d60ad1e9f0026d3c65cfc315a702541f09522b120be9b23"} Feb 20 07:00:50 crc kubenswrapper[4492]: I0220 07:00:50.079872 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18f922e8-15c1-49a7-af1b-38a85b924404","Type":"ContainerStarted","Data":"1bfa22cdabf654b47ef25aa38e86d86349c82238fdabd90372c4450a5a2512be"} Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.532756 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.618688 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d8ddddc7-m6b6t"] Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.619004 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" podUID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerName="dnsmasq-dns" containerID="cri-o://e1d88bafc59d2a7026a983fdf0fbaf0d489bae206d1f432191e15b7c691d0d26" gracePeriod=10 Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.812583 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c759b9c5-fr6n5"] Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.814993 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.876112 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c759b9c5-fr6n5"] Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.939406 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-ovsdbserver-sb\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.939534 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-dns-svc\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.939614 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-dns-swift-storage-0\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.939655 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4rpj\" (UniqueName: \"kubernetes.io/projected/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-kube-api-access-x4rpj\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.939687 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-openstack-edpm-ipam\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.939708 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-ovsdbserver-nb\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:55 crc kubenswrapper[4492]: I0220 07:00:55.939742 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-config\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.042099 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-dns-svc\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.042170 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-dns-swift-storage-0\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.042219 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4rpj\" (UniqueName: \"kubernetes.io/projected/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-kube-api-access-x4rpj\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.042253 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-openstack-edpm-ipam\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.042287 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-ovsdbserver-nb\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.042325 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-config\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.042398 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-ovsdbserver-sb\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.043189 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-ovsdbserver-sb\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.043755 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-dns-svc\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.045250 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-ovsdbserver-nb\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.045315 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-config\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.046111 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-dns-swift-storage-0\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.046637 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-openstack-edpm-ipam\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.069234 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4rpj\" (UniqueName: \"kubernetes.io/projected/b70a74af-cdf1-4c6a-92ce-1148516ae7ee-kube-api-access-x4rpj\") pod \"dnsmasq-dns-7c759b9c5-fr6n5\" (UID: \"b70a74af-cdf1-4c6a-92ce-1148516ae7ee\") " pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.130735 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.174017 4492 generic.go:334] "Generic (PLEG): container finished" podID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerID="e1d88bafc59d2a7026a983fdf0fbaf0d489bae206d1f432191e15b7c691d0d26" exitCode=0 Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.174294 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" event={"ID":"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372","Type":"ContainerDied","Data":"e1d88bafc59d2a7026a983fdf0fbaf0d489bae206d1f432191e15b7c691d0d26"} Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.174324 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" event={"ID":"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372","Type":"ContainerDied","Data":"fd36cc307bc99820412a68a6b2e7521260fea98cfa49a7e66d4241f1626d6a9d"} Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.174339 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd36cc307bc99820412a68a6b2e7521260fea98cfa49a7e66d4241f1626d6a9d" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.253313 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.349293 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxq6k\" (UniqueName: \"kubernetes.io/projected/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-kube-api-access-cxq6k\") pod \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.349397 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-swift-storage-0\") pod \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.349666 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-svc\") pod \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.349695 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-config\") pod \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.349751 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-nb\") pod \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.349830 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-sb\") pod \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\" (UID: \"ca34c1ba-31a4-43f6-9fc4-3ca9d978c372\") " Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.357271 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-kube-api-access-cxq6k" (OuterVolumeSpecName: "kube-api-access-cxq6k") pod "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" (UID: "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372"). InnerVolumeSpecName "kube-api-access-cxq6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.401318 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-config" (OuterVolumeSpecName: "config") pod "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" (UID: "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.402951 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" (UID: "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.408836 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" (UID: "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.415626 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" (UID: "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.429186 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" (UID: "ca34c1ba-31a4-43f6-9fc4-3ca9d978c372"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.453101 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxq6k\" (UniqueName: \"kubernetes.io/projected/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-kube-api-access-cxq6k\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.453132 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.453143 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.453153 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-config\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.453161 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.453169 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 07:00:56 crc kubenswrapper[4492]: I0220 07:00:56.613364 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c759b9c5-fr6n5"] Feb 20 07:00:57 crc kubenswrapper[4492]: I0220 07:00:57.188739 4492 generic.go:334] "Generic (PLEG): container finished" podID="b70a74af-cdf1-4c6a-92ce-1148516ae7ee" containerID="59963af43bee30adf53bb62517115dfc55f346a9487c27419bbeb4e60f617a8e" exitCode=0 Feb 20 07:00:57 crc kubenswrapper[4492]: I0220 07:00:57.188882 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65d8ddddc7-m6b6t" Feb 20 07:00:57 crc kubenswrapper[4492]: I0220 07:00:57.190757 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" event={"ID":"b70a74af-cdf1-4c6a-92ce-1148516ae7ee","Type":"ContainerDied","Data":"59963af43bee30adf53bb62517115dfc55f346a9487c27419bbeb4e60f617a8e"} Feb 20 07:00:57 crc kubenswrapper[4492]: I0220 07:00:57.190830 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" event={"ID":"b70a74af-cdf1-4c6a-92ce-1148516ae7ee","Type":"ContainerStarted","Data":"a207f29f94d8a7eed3e9b23d62dede9915c2914847e4bd363cbe3ea173ba978a"} Feb 20 07:00:57 crc kubenswrapper[4492]: I0220 07:00:57.301330 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65d8ddddc7-m6b6t"] Feb 20 07:00:57 crc kubenswrapper[4492]: I0220 07:00:57.311068 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65d8ddddc7-m6b6t"] Feb 20 07:00:57 crc kubenswrapper[4492]: I0220 07:00:57.570182 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" path="/var/lib/kubelet/pods/ca34c1ba-31a4-43f6-9fc4-3ca9d978c372/volumes" Feb 20 07:00:58 crc kubenswrapper[4492]: I0220 07:00:58.202761 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" event={"ID":"b70a74af-cdf1-4c6a-92ce-1148516ae7ee","Type":"ContainerStarted","Data":"466473fd0653ae137d22620cd5db38d722ab880c0a61a69875be2a898b319c85"} Feb 20 07:00:58 crc kubenswrapper[4492]: I0220 07:00:58.203268 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:00:58 crc kubenswrapper[4492]: I0220 07:00:58.225868 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" podStartSLOduration=3.225844855 podStartE2EDuration="3.225844855s" podCreationTimestamp="2026-02-20 07:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:00:58.220855283 +0000 UTC m=+1214.992144261" watchObservedRunningTime="2026-02-20 07:00:58.225844855 +0000 UTC m=+1214.997133834" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.146787 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29526181-rdldl"] Feb 20 07:01:00 crc kubenswrapper[4492]: E0220 07:01:00.147880 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerName="dnsmasq-dns" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.147914 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerName="dnsmasq-dns" Feb 20 07:01:00 crc kubenswrapper[4492]: E0220 07:01:00.147958 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerName="init" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.147968 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerName="init" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.148247 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca34c1ba-31a4-43f6-9fc4-3ca9d978c372" containerName="dnsmasq-dns" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.149319 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.157269 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526181-rdldl"] Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.257980 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-config-data\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.258313 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxxxj\" (UniqueName: \"kubernetes.io/projected/40796381-0eec-4c7d-b1f0-a24b513ea411-kube-api-access-hxxxj\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.258366 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-fernet-keys\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.258541 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-combined-ca-bundle\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.361624 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxxxj\" (UniqueName: \"kubernetes.io/projected/40796381-0eec-4c7d-b1f0-a24b513ea411-kube-api-access-hxxxj\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.361673 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-fernet-keys\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.361752 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-combined-ca-bundle\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.362077 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-config-data\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.369453 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-config-data\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.371337 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-combined-ca-bundle\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.372045 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-fernet-keys\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.378029 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxxxj\" (UniqueName: \"kubernetes.io/projected/40796381-0eec-4c7d-b1f0-a24b513ea411-kube-api-access-hxxxj\") pod \"keystone-cron-29526181-rdldl\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.485281 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:00 crc kubenswrapper[4492]: I0220 07:01:00.907820 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526181-rdldl"] Feb 20 07:01:01 crc kubenswrapper[4492]: I0220 07:01:01.231792 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526181-rdldl" event={"ID":"40796381-0eec-4c7d-b1f0-a24b513ea411","Type":"ContainerStarted","Data":"a3f24a2e16a18487fdd8eea6f8755a1334ad7173f9f1caed4d26f4072a9ae3b7"} Feb 20 07:01:01 crc kubenswrapper[4492]: I0220 07:01:01.231864 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526181-rdldl" event={"ID":"40796381-0eec-4c7d-b1f0-a24b513ea411","Type":"ContainerStarted","Data":"319ca0f0e7d1256a85822fcf519dcf00e80f742a3b29b45514ac6698782d3d44"} Feb 20 07:01:01 crc kubenswrapper[4492]: I0220 07:01:01.258288 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29526181-rdldl" podStartSLOduration=1.258263388 podStartE2EDuration="1.258263388s" podCreationTimestamp="2026-02-20 07:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:01:01.25315449 +0000 UTC m=+1218.024443468" watchObservedRunningTime="2026-02-20 07:01:01.258263388 +0000 UTC m=+1218.029552366" Feb 20 07:01:03 crc kubenswrapper[4492]: I0220 07:01:03.862046 4492 scope.go:117] "RemoveContainer" containerID="36ddd304e22b8b317e53833f8714a3f92f0c978cd473986cfd86e344942523b1" Feb 20 07:01:04 crc kubenswrapper[4492]: I0220 07:01:04.267044 4492 generic.go:334] "Generic (PLEG): container finished" podID="40796381-0eec-4c7d-b1f0-a24b513ea411" containerID="a3f24a2e16a18487fdd8eea6f8755a1334ad7173f9f1caed4d26f4072a9ae3b7" exitCode=0 Feb 20 07:01:04 crc kubenswrapper[4492]: I0220 07:01:04.267136 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526181-rdldl" event={"ID":"40796381-0eec-4c7d-b1f0-a24b513ea411","Type":"ContainerDied","Data":"a3f24a2e16a18487fdd8eea6f8755a1334ad7173f9f1caed4d26f4072a9ae3b7"} Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.551379 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.696836 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxxxj\" (UniqueName: \"kubernetes.io/projected/40796381-0eec-4c7d-b1f0-a24b513ea411-kube-api-access-hxxxj\") pod \"40796381-0eec-4c7d-b1f0-a24b513ea411\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.697055 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-combined-ca-bundle\") pod \"40796381-0eec-4c7d-b1f0-a24b513ea411\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.697093 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-config-data\") pod \"40796381-0eec-4c7d-b1f0-a24b513ea411\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.697111 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-fernet-keys\") pod \"40796381-0eec-4c7d-b1f0-a24b513ea411\" (UID: \"40796381-0eec-4c7d-b1f0-a24b513ea411\") " Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.727615 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "40796381-0eec-4c7d-b1f0-a24b513ea411" (UID: "40796381-0eec-4c7d-b1f0-a24b513ea411"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.732611 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40796381-0eec-4c7d-b1f0-a24b513ea411-kube-api-access-hxxxj" (OuterVolumeSpecName: "kube-api-access-hxxxj") pod "40796381-0eec-4c7d-b1f0-a24b513ea411" (UID: "40796381-0eec-4c7d-b1f0-a24b513ea411"). InnerVolumeSpecName "kube-api-access-hxxxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.739225 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40796381-0eec-4c7d-b1f0-a24b513ea411" (UID: "40796381-0eec-4c7d-b1f0-a24b513ea411"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.763481 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-config-data" (OuterVolumeSpecName: "config-data") pod "40796381-0eec-4c7d-b1f0-a24b513ea411" (UID: "40796381-0eec-4c7d-b1f0-a24b513ea411"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.800143 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.800177 4492 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.800187 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40796381-0eec-4c7d-b1f0-a24b513ea411-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:05 crc kubenswrapper[4492]: I0220 07:01:05.800197 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxxxj\" (UniqueName: \"kubernetes.io/projected/40796381-0eec-4c7d-b1f0-a24b513ea411-kube-api-access-hxxxj\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.132963 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c759b9c5-fr6n5" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.202377 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57486d48df-tv7hf"] Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.202732 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" podUID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerName="dnsmasq-dns" containerID="cri-o://1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491" gracePeriod=10 Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.287928 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526181-rdldl" event={"ID":"40796381-0eec-4c7d-b1f0-a24b513ea411","Type":"ContainerDied","Data":"319ca0f0e7d1256a85822fcf519dcf00e80f742a3b29b45514ac6698782d3d44"} Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.287968 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="319ca0f0e7d1256a85822fcf519dcf00e80f742a3b29b45514ac6698782d3d44" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.288023 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526181-rdldl" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.648888 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.831764 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-nb\") pod \"f7408165-ed60-4ff1-a838-a5e98cd2710f\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.832230 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-swift-storage-0\") pod \"f7408165-ed60-4ff1-a838-a5e98cd2710f\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.832261 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2dfm\" (UniqueName: \"kubernetes.io/projected/f7408165-ed60-4ff1-a838-a5e98cd2710f-kube-api-access-q2dfm\") pod \"f7408165-ed60-4ff1-a838-a5e98cd2710f\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.832655 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-openstack-edpm-ipam\") pod \"f7408165-ed60-4ff1-a838-a5e98cd2710f\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.832711 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-svc\") pod \"f7408165-ed60-4ff1-a838-a5e98cd2710f\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.832752 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-sb\") pod \"f7408165-ed60-4ff1-a838-a5e98cd2710f\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.832911 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-config\") pod \"f7408165-ed60-4ff1-a838-a5e98cd2710f\" (UID: \"f7408165-ed60-4ff1-a838-a5e98cd2710f\") " Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.840870 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7408165-ed60-4ff1-a838-a5e98cd2710f-kube-api-access-q2dfm" (OuterVolumeSpecName: "kube-api-access-q2dfm") pod "f7408165-ed60-4ff1-a838-a5e98cd2710f" (UID: "f7408165-ed60-4ff1-a838-a5e98cd2710f"). InnerVolumeSpecName "kube-api-access-q2dfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.879363 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "f7408165-ed60-4ff1-a838-a5e98cd2710f" (UID: "f7408165-ed60-4ff1-a838-a5e98cd2710f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.883160 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-config" (OuterVolumeSpecName: "config") pod "f7408165-ed60-4ff1-a838-a5e98cd2710f" (UID: "f7408165-ed60-4ff1-a838-a5e98cd2710f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.887810 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f7408165-ed60-4ff1-a838-a5e98cd2710f" (UID: "f7408165-ed60-4ff1-a838-a5e98cd2710f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.896653 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f7408165-ed60-4ff1-a838-a5e98cd2710f" (UID: "f7408165-ed60-4ff1-a838-a5e98cd2710f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.898190 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f7408165-ed60-4ff1-a838-a5e98cd2710f" (UID: "f7408165-ed60-4ff1-a838-a5e98cd2710f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.901540 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f7408165-ed60-4ff1-a838-a5e98cd2710f" (UID: "f7408165-ed60-4ff1-a838-a5e98cd2710f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.936807 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.936835 4492 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.936848 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2dfm\" (UniqueName: \"kubernetes.io/projected/f7408165-ed60-4ff1-a838-a5e98cd2710f-kube-api-access-q2dfm\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.936860 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.936870 4492 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.936897 4492 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:06 crc kubenswrapper[4492]: I0220 07:01:06.936906 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7408165-ed60-4ff1-a838-a5e98cd2710f-config\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.298616 4492 generic.go:334] "Generic (PLEG): container finished" podID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerID="1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491" exitCode=0 Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.298670 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" event={"ID":"f7408165-ed60-4ff1-a838-a5e98cd2710f","Type":"ContainerDied","Data":"1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491"} Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.298699 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.298720 4492 scope.go:117] "RemoveContainer" containerID="1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.298705 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57486d48df-tv7hf" event={"ID":"f7408165-ed60-4ff1-a838-a5e98cd2710f","Type":"ContainerDied","Data":"3ca7f5a72ad5764b76696bce9a6df90793eed5f5364af549548f61279ef10daa"} Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.319895 4492 scope.go:117] "RemoveContainer" containerID="75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.325668 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57486d48df-tv7hf"] Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.336875 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57486d48df-tv7hf"] Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.341555 4492 scope.go:117] "RemoveContainer" containerID="1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491" Feb 20 07:01:07 crc kubenswrapper[4492]: E0220 07:01:07.342032 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491\": container with ID starting with 1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491 not found: ID does not exist" containerID="1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.342062 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491"} err="failed to get container status \"1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491\": rpc error: code = NotFound desc = could not find container \"1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491\": container with ID starting with 1aceed10c0f414389f7d6d8415a80c0e950a4cecba78bf397b6923192bb08491 not found: ID does not exist" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.342084 4492 scope.go:117] "RemoveContainer" containerID="75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a" Feb 20 07:01:07 crc kubenswrapper[4492]: E0220 07:01:07.342308 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a\": container with ID starting with 75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a not found: ID does not exist" containerID="75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.342329 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a"} err="failed to get container status \"75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a\": rpc error: code = NotFound desc = could not find container \"75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a\": container with ID starting with 75e23b1d71bb58b6235ad3c582f38ca0a1a2194240058ab5c939c3090665047a not found: ID does not exist" Feb 20 07:01:07 crc kubenswrapper[4492]: I0220 07:01:07.567077 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7408165-ed60-4ff1-a838-a5e98cd2710f" path="/var/lib/kubelet/pods/f7408165-ed60-4ff1-a838-a5e98cd2710f/volumes" Feb 20 07:01:20 crc kubenswrapper[4492]: I0220 07:01:20.461910 4492 generic.go:334] "Generic (PLEG): container finished" podID="a4dfa138-8cc8-40e0-b2df-21449a994146" containerID="5cdc5ebab8d08f295d60ad1e9f0026d3c65cfc315a702541f09522b120be9b23" exitCode=0 Feb 20 07:01:20 crc kubenswrapper[4492]: I0220 07:01:20.461992 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a4dfa138-8cc8-40e0-b2df-21449a994146","Type":"ContainerDied","Data":"5cdc5ebab8d08f295d60ad1e9f0026d3c65cfc315a702541f09522b120be9b23"} Feb 20 07:01:21 crc kubenswrapper[4492]: I0220 07:01:21.479354 4492 generic.go:334] "Generic (PLEG): container finished" podID="18f922e8-15c1-49a7-af1b-38a85b924404" containerID="1bfa22cdabf654b47ef25aa38e86d86349c82238fdabd90372c4450a5a2512be" exitCode=0 Feb 20 07:01:21 crc kubenswrapper[4492]: I0220 07:01:21.479445 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18f922e8-15c1-49a7-af1b-38a85b924404","Type":"ContainerDied","Data":"1bfa22cdabf654b47ef25aa38e86d86349c82238fdabd90372c4450a5a2512be"} Feb 20 07:01:21 crc kubenswrapper[4492]: I0220 07:01:21.483914 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a4dfa138-8cc8-40e0-b2df-21449a994146","Type":"ContainerStarted","Data":"39c7949d279d21744697a4091f502e592b8d322aba9053f03ad9c60c340ab1a4"} Feb 20 07:01:21 crc kubenswrapper[4492]: I0220 07:01:21.484439 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 20 07:01:21 crc kubenswrapper[4492]: I0220 07:01:21.561102 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.561012255 podStartE2EDuration="35.561012255s" podCreationTimestamp="2026-02-20 07:00:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:01:21.535025963 +0000 UTC m=+1238.306314932" watchObservedRunningTime="2026-02-20 07:01:21.561012255 +0000 UTC m=+1238.332301233" Feb 20 07:01:22 crc kubenswrapper[4492]: I0220 07:01:22.497587 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"18f922e8-15c1-49a7-af1b-38a85b924404","Type":"ContainerStarted","Data":"834e07b0865ca546238324eadf732e5e69b7a0516f091aaa5372994e271aa416"} Feb 20 07:01:22 crc kubenswrapper[4492]: I0220 07:01:22.498259 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:01:22 crc kubenswrapper[4492]: I0220 07:01:22.527957 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.527932704 podStartE2EDuration="35.527932704s" podCreationTimestamp="2026-02-20 07:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:01:22.518712438 +0000 UTC m=+1239.290001416" watchObservedRunningTime="2026-02-20 07:01:22.527932704 +0000 UTC m=+1239.299221681" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.163908 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd"] Feb 20 07:01:24 crc kubenswrapper[4492]: E0220 07:01:24.164369 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerName="init" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.164385 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerName="init" Feb 20 07:01:24 crc kubenswrapper[4492]: E0220 07:01:24.164395 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40796381-0eec-4c7d-b1f0-a24b513ea411" containerName="keystone-cron" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.164401 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="40796381-0eec-4c7d-b1f0-a24b513ea411" containerName="keystone-cron" Feb 20 07:01:24 crc kubenswrapper[4492]: E0220 07:01:24.164411 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerName="dnsmasq-dns" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.164418 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerName="dnsmasq-dns" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.164653 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7408165-ed60-4ff1-a838-a5e98cd2710f" containerName="dnsmasq-dns" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.164682 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="40796381-0eec-4c7d-b1f0-a24b513ea411" containerName="keystone-cron" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.165316 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.172288 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.172393 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.172503 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.173183 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.181306 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd"] Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.252298 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.252447 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6r9f\" (UniqueName: \"kubernetes.io/projected/a54176de-a2a7-4819-aa62-f5da167b49c0-kube-api-access-j6r9f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.252526 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.252667 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.354437 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.354537 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.354649 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6r9f\" (UniqueName: \"kubernetes.io/projected/a54176de-a2a7-4819-aa62-f5da167b49c0-kube-api-access-j6r9f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.354757 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.363026 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.365709 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.375355 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.378422 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6r9f\" (UniqueName: \"kubernetes.io/projected/a54176de-a2a7-4819-aa62-f5da167b49c0-kube-api-access-j6r9f\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:24 crc kubenswrapper[4492]: I0220 07:01:24.483556 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:25 crc kubenswrapper[4492]: I0220 07:01:25.127193 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd"] Feb 20 07:01:25 crc kubenswrapper[4492]: I0220 07:01:25.550178 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" event={"ID":"a54176de-a2a7-4819-aa62-f5da167b49c0","Type":"ContainerStarted","Data":"82b2dfe886f6f0d8c82596d98a711679163849865a2c084841754c230cbf0eb0"} Feb 20 07:01:36 crc kubenswrapper[4492]: I0220 07:01:36.434661 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 20 07:01:36 crc kubenswrapper[4492]: I0220 07:01:36.675008 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" event={"ID":"a54176de-a2a7-4819-aa62-f5da167b49c0","Type":"ContainerStarted","Data":"4b9810413ab8606baf46086ef06bd18519a303219a98a8c119d2ebea3bbfca0c"} Feb 20 07:01:36 crc kubenswrapper[4492]: I0220 07:01:36.691058 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" podStartSLOduration=2.270811855 podStartE2EDuration="12.691023558s" podCreationTimestamp="2026-02-20 07:01:24 +0000 UTC" firstStartedPulling="2026-02-20 07:01:25.142156609 +0000 UTC m=+1241.913445587" lastFinishedPulling="2026-02-20 07:01:35.562368312 +0000 UTC m=+1252.333657290" observedRunningTime="2026-02-20 07:01:36.687913227 +0000 UTC m=+1253.459202205" watchObservedRunningTime="2026-02-20 07:01:36.691023558 +0000 UTC m=+1253.462312536" Feb 20 07:01:37 crc kubenswrapper[4492]: I0220 07:01:37.753771 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 20 07:01:47 crc kubenswrapper[4492]: I0220 07:01:47.805030 4492 generic.go:334] "Generic (PLEG): container finished" podID="a54176de-a2a7-4819-aa62-f5da167b49c0" containerID="4b9810413ab8606baf46086ef06bd18519a303219a98a8c119d2ebea3bbfca0c" exitCode=0 Feb 20 07:01:47 crc kubenswrapper[4492]: I0220 07:01:47.805091 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" event={"ID":"a54176de-a2a7-4819-aa62-f5da167b49c0","Type":"ContainerDied","Data":"4b9810413ab8606baf46086ef06bd18519a303219a98a8c119d2ebea3bbfca0c"} Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.325143 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.490513 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-inventory\") pod \"a54176de-a2a7-4819-aa62-f5da167b49c0\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.490686 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-ssh-key-openstack-edpm-ipam\") pod \"a54176de-a2a7-4819-aa62-f5da167b49c0\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.490720 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-repo-setup-combined-ca-bundle\") pod \"a54176de-a2a7-4819-aa62-f5da167b49c0\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.490844 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6r9f\" (UniqueName: \"kubernetes.io/projected/a54176de-a2a7-4819-aa62-f5da167b49c0-kube-api-access-j6r9f\") pod \"a54176de-a2a7-4819-aa62-f5da167b49c0\" (UID: \"a54176de-a2a7-4819-aa62-f5da167b49c0\") " Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.499584 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a54176de-a2a7-4819-aa62-f5da167b49c0" (UID: "a54176de-a2a7-4819-aa62-f5da167b49c0"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.501813 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54176de-a2a7-4819-aa62-f5da167b49c0-kube-api-access-j6r9f" (OuterVolumeSpecName: "kube-api-access-j6r9f") pod "a54176de-a2a7-4819-aa62-f5da167b49c0" (UID: "a54176de-a2a7-4819-aa62-f5da167b49c0"). InnerVolumeSpecName "kube-api-access-j6r9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.528599 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-inventory" (OuterVolumeSpecName: "inventory") pod "a54176de-a2a7-4819-aa62-f5da167b49c0" (UID: "a54176de-a2a7-4819-aa62-f5da167b49c0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.528997 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a54176de-a2a7-4819-aa62-f5da167b49c0" (UID: "a54176de-a2a7-4819-aa62-f5da167b49c0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.596145 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.596181 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.596194 4492 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a54176de-a2a7-4819-aa62-f5da167b49c0-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.596207 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6r9f\" (UniqueName: \"kubernetes.io/projected/a54176de-a2a7-4819-aa62-f5da167b49c0-kube-api-access-j6r9f\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.828622 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" event={"ID":"a54176de-a2a7-4819-aa62-f5da167b49c0","Type":"ContainerDied","Data":"82b2dfe886f6f0d8c82596d98a711679163849865a2c084841754c230cbf0eb0"} Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.828681 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82b2dfe886f6f0d8c82596d98a711679163849865a2c084841754c230cbf0eb0" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.828715 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.900727 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg"] Feb 20 07:01:49 crc kubenswrapper[4492]: E0220 07:01:49.901648 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54176de-a2a7-4819-aa62-f5da167b49c0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.901685 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54176de-a2a7-4819-aa62-f5da167b49c0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.902044 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54176de-a2a7-4819-aa62-f5da167b49c0" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.903172 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.910143 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.910955 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.911079 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.912606 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:01:49 crc kubenswrapper[4492]: I0220 07:01:49.921265 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg"] Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.007889 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.008830 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.009065 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnv9r\" (UniqueName: \"kubernetes.io/projected/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-kube-api-access-vnv9r\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.112568 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.112960 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.113099 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnv9r\" (UniqueName: \"kubernetes.io/projected/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-kube-api-access-vnv9r\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.117988 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.117997 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.129621 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnv9r\" (UniqueName: \"kubernetes.io/projected/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-kube-api-access-vnv9r\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qgbrg\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.223954 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.748299 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg"] Feb 20 07:01:50 crc kubenswrapper[4492]: I0220 07:01:50.844466 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" event={"ID":"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793","Type":"ContainerStarted","Data":"7999ee96b6fd31b8fcf8298ad6b9368dbf7a6c7543ad1563bdeeb27008b61d52"} Feb 20 07:01:51 crc kubenswrapper[4492]: I0220 07:01:51.860130 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" event={"ID":"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793","Type":"ContainerStarted","Data":"7cf4d08a206acafcbc5165be26d2183c9a29f45609c52d53f6feb77007e66861"} Feb 20 07:01:51 crc kubenswrapper[4492]: I0220 07:01:51.892483 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" podStartSLOduration=2.358890051 podStartE2EDuration="2.892447979s" podCreationTimestamp="2026-02-20 07:01:49 +0000 UTC" firstStartedPulling="2026-02-20 07:01:50.757485447 +0000 UTC m=+1267.528774425" lastFinishedPulling="2026-02-20 07:01:51.291043375 +0000 UTC m=+1268.062332353" observedRunningTime="2026-02-20 07:01:51.880743333 +0000 UTC m=+1268.652032310" watchObservedRunningTime="2026-02-20 07:01:51.892447979 +0000 UTC m=+1268.663736957" Feb 20 07:01:53 crc kubenswrapper[4492]: I0220 07:01:53.882759 4492 generic.go:334] "Generic (PLEG): container finished" podID="d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793" containerID="7cf4d08a206acafcbc5165be26d2183c9a29f45609c52d53f6feb77007e66861" exitCode=0 Feb 20 07:01:53 crc kubenswrapper[4492]: I0220 07:01:53.883158 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" event={"ID":"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793","Type":"ContainerDied","Data":"7cf4d08a206acafcbc5165be26d2183c9a29f45609c52d53f6feb77007e66861"} Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.251453 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.445898 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-inventory\") pod \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.446058 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnv9r\" (UniqueName: \"kubernetes.io/projected/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-kube-api-access-vnv9r\") pod \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.446109 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-ssh-key-openstack-edpm-ipam\") pod \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\" (UID: \"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793\") " Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.475756 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-kube-api-access-vnv9r" (OuterVolumeSpecName: "kube-api-access-vnv9r") pod "d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793" (UID: "d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793"). InnerVolumeSpecName "kube-api-access-vnv9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.488772 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-inventory" (OuterVolumeSpecName: "inventory") pod "d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793" (UID: "d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.489313 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793" (UID: "d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.549446 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.549500 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnv9r\" (UniqueName: \"kubernetes.io/projected/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-kube-api-access-vnv9r\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.549515 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.902265 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" event={"ID":"d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793","Type":"ContainerDied","Data":"7999ee96b6fd31b8fcf8298ad6b9368dbf7a6c7543ad1563bdeeb27008b61d52"} Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.902309 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7999ee96b6fd31b8fcf8298ad6b9368dbf7a6c7543ad1563bdeeb27008b61d52" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.902312 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qgbrg" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.959314 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4"] Feb 20 07:01:55 crc kubenswrapper[4492]: E0220 07:01:55.959804 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.959825 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.960013 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.960742 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.963195 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.963395 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.963658 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.963770 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:01:55 crc kubenswrapper[4492]: I0220 07:01:55.984328 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4"] Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.060187 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.060262 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df2xs\" (UniqueName: \"kubernetes.io/projected/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-kube-api-access-df2xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.060338 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.060651 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.163738 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.164088 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df2xs\" (UniqueName: \"kubernetes.io/projected/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-kube-api-access-df2xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.164171 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.164495 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.177702 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.179463 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.200261 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.201592 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df2xs\" (UniqueName: \"kubernetes.io/projected/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-kube-api-access-df2xs\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.277692 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.865011 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4"] Feb 20 07:01:56 crc kubenswrapper[4492]: I0220 07:01:56.920407 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" event={"ID":"52acdbf4-c839-4ed1-bbd7-a7674e2bab77","Type":"ContainerStarted","Data":"5185eac91cf2132dd4569b59098596b01c4106a707c95450d9018b64f576d1aa"} Feb 20 07:01:57 crc kubenswrapper[4492]: I0220 07:01:57.934760 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" event={"ID":"52acdbf4-c839-4ed1-bbd7-a7674e2bab77","Type":"ContainerStarted","Data":"a9f411ef312d25a6c9fb10b354ca1e7f3ed182a47dfda129f17021594a8e374b"} Feb 20 07:01:57 crc kubenswrapper[4492]: I0220 07:01:57.962766 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" podStartSLOduration=2.502203673 podStartE2EDuration="2.96274888s" podCreationTimestamp="2026-02-20 07:01:55 +0000 UTC" firstStartedPulling="2026-02-20 07:01:56.872996896 +0000 UTC m=+1273.644285874" lastFinishedPulling="2026-02-20 07:01:57.333542102 +0000 UTC m=+1274.104831081" observedRunningTime="2026-02-20 07:01:57.950322994 +0000 UTC m=+1274.721611972" watchObservedRunningTime="2026-02-20 07:01:57.96274888 +0000 UTC m=+1274.734037859" Feb 20 07:02:03 crc kubenswrapper[4492]: I0220 07:02:03.981441 4492 scope.go:117] "RemoveContainer" containerID="a7281b184eed81c267362a89b0acaa26c369b789487bef03cf8be5de82cd79fb" Feb 20 07:02:09 crc kubenswrapper[4492]: I0220 07:02:09.310852 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:02:09 crc kubenswrapper[4492]: I0220 07:02:09.311410 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:02:39 crc kubenswrapper[4492]: I0220 07:02:39.311314 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:02:39 crc kubenswrapper[4492]: I0220 07:02:39.312002 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:03:04 crc kubenswrapper[4492]: I0220 07:03:04.069559 4492 scope.go:117] "RemoveContainer" containerID="bcabc6f3f996af9ef5ee90c40113ce781e659a7a4c4b6a4e9ab2297a0db6c2a1" Feb 20 07:03:04 crc kubenswrapper[4492]: I0220 07:03:04.130462 4492 scope.go:117] "RemoveContainer" containerID="da1f81c37dc29e26c561d18931d35d6667a0d9062ffa4f5ae9bc3c86d436dd0c" Feb 20 07:03:04 crc kubenswrapper[4492]: I0220 07:03:04.168294 4492 scope.go:117] "RemoveContainer" containerID="147dc48098e65dcc3fa6a521780d7102c598513b72fbd316186df5c77c6972b1" Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.311423 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.311861 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.311931 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.312626 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ab8284c72f81bddf931f4abee5f04c18596ed9e487d4c6a5d3e615ec918dcbba"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.312678 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://ab8284c72f81bddf931f4abee5f04c18596ed9e487d4c6a5d3e615ec918dcbba" gracePeriod=600 Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.684846 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="ab8284c72f81bddf931f4abee5f04c18596ed9e487d4c6a5d3e615ec918dcbba" exitCode=0 Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.684919 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"ab8284c72f81bddf931f4abee5f04c18596ed9e487d4c6a5d3e615ec918dcbba"} Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.685167 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe"} Feb 20 07:03:09 crc kubenswrapper[4492]: I0220 07:03:09.685208 4492 scope.go:117] "RemoveContainer" containerID="1968b24edc23a5de4ecabc4dcc18c03f061c8965af1f7e77792ae0d80771d5c6" Feb 20 07:04:04 crc kubenswrapper[4492]: I0220 07:04:04.264406 4492 scope.go:117] "RemoveContainer" containerID="c66c7d455e6be216939439401d091ce8dd637fd073bd71c7aa443efca31619be" Feb 20 07:04:04 crc kubenswrapper[4492]: I0220 07:04:04.290672 4492 scope.go:117] "RemoveContainer" containerID="23b25f0ddbab7f6803ecc13c5e0c9b0aebf22d7f3c1bf84aac2dcc3ba90297e1" Feb 20 07:04:04 crc kubenswrapper[4492]: I0220 07:04:04.314038 4492 scope.go:117] "RemoveContainer" containerID="6b658b53be664062d0337d6b6fc83749327873afca9013f7a0a14b9450478cae" Feb 20 07:04:04 crc kubenswrapper[4492]: I0220 07:04:04.335752 4492 scope.go:117] "RemoveContainer" containerID="288368ed8b45752696739a4b85c3bcd829998eec79d3edd3d7ed060f74d30e08" Feb 20 07:04:04 crc kubenswrapper[4492]: I0220 07:04:04.354840 4492 scope.go:117] "RemoveContainer" containerID="8af0e4b9b8c0ec589978608ba21391247e42f776ce4089b3eeacaa4bf774448e" Feb 20 07:04:50 crc kubenswrapper[4492]: I0220 07:04:50.802628 4492 generic.go:334] "Generic (PLEG): container finished" podID="52acdbf4-c839-4ed1-bbd7-a7674e2bab77" containerID="a9f411ef312d25a6c9fb10b354ca1e7f3ed182a47dfda129f17021594a8e374b" exitCode=0 Feb 20 07:04:50 crc kubenswrapper[4492]: I0220 07:04:50.802705 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" event={"ID":"52acdbf4-c839-4ed1-bbd7-a7674e2bab77","Type":"ContainerDied","Data":"a9f411ef312d25a6c9fb10b354ca1e7f3ed182a47dfda129f17021594a8e374b"} Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.176269 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.223647 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-bootstrap-combined-ca-bundle\") pod \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.223693 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-ssh-key-openstack-edpm-ipam\") pod \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.223986 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-inventory\") pod \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.224071 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df2xs\" (UniqueName: \"kubernetes.io/projected/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-kube-api-access-df2xs\") pod \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\" (UID: \"52acdbf4-c839-4ed1-bbd7-a7674e2bab77\") " Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.233200 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "52acdbf4-c839-4ed1-bbd7-a7674e2bab77" (UID: "52acdbf4-c839-4ed1-bbd7-a7674e2bab77"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.234218 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-kube-api-access-df2xs" (OuterVolumeSpecName: "kube-api-access-df2xs") pod "52acdbf4-c839-4ed1-bbd7-a7674e2bab77" (UID: "52acdbf4-c839-4ed1-bbd7-a7674e2bab77"). InnerVolumeSpecName "kube-api-access-df2xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.247669 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "52acdbf4-c839-4ed1-bbd7-a7674e2bab77" (UID: "52acdbf4-c839-4ed1-bbd7-a7674e2bab77"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.252575 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-inventory" (OuterVolumeSpecName: "inventory") pod "52acdbf4-c839-4ed1-bbd7-a7674e2bab77" (UID: "52acdbf4-c839-4ed1-bbd7-a7674e2bab77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.328194 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df2xs\" (UniqueName: \"kubernetes.io/projected/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-kube-api-access-df2xs\") on node \"crc\" DevicePath \"\"" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.328230 4492 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.328245 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.328259 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/52acdbf4-c839-4ed1-bbd7-a7674e2bab77-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.840327 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" event={"ID":"52acdbf4-c839-4ed1-bbd7-a7674e2bab77","Type":"ContainerDied","Data":"5185eac91cf2132dd4569b59098596b01c4106a707c95450d9018b64f576d1aa"} Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.840657 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5185eac91cf2132dd4569b59098596b01c4106a707c95450d9018b64f576d1aa" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.840385 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.904035 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr"] Feb 20 07:04:52 crc kubenswrapper[4492]: E0220 07:04:52.904438 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52acdbf4-c839-4ed1-bbd7-a7674e2bab77" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.904456 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="52acdbf4-c839-4ed1-bbd7-a7674e2bab77" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.904682 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="52acdbf4-c839-4ed1-bbd7-a7674e2bab77" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.905356 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.912800 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.913174 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.913254 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.913276 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.916430 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr"] Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.939335 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.939419 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grgn8\" (UniqueName: \"kubernetes.io/projected/d0c74792-a006-4752-8403-70947f6fec63-kube-api-access-grgn8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:52 crc kubenswrapper[4492]: I0220 07:04:52.939483 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.042008 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.042216 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grgn8\" (UniqueName: \"kubernetes.io/projected/d0c74792-a006-4752-8403-70947f6fec63-kube-api-access-grgn8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.042339 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.049511 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.049564 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.064274 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grgn8\" (UniqueName: \"kubernetes.io/projected/d0c74792-a006-4752-8403-70947f6fec63-kube-api-access-grgn8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.218755 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.719947 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr"] Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.721214 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:04:53 crc kubenswrapper[4492]: I0220 07:04:53.852319 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" event={"ID":"d0c74792-a006-4752-8403-70947f6fec63","Type":"ContainerStarted","Data":"e5b1abd96c168e366c2eb5a9f50873a21531055dd6d81abb5b77cd12b17ce048"} Feb 20 07:04:54 crc kubenswrapper[4492]: I0220 07:04:54.866144 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" event={"ID":"d0c74792-a006-4752-8403-70947f6fec63","Type":"ContainerStarted","Data":"9847b6eca3297f0d076cfef8bd706cd06a64f259904ab9fb7f1f54bbfaeb4db3"} Feb 20 07:04:54 crc kubenswrapper[4492]: I0220 07:04:54.885793 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" podStartSLOduration=2.35469461 podStartE2EDuration="2.885770607s" podCreationTimestamp="2026-02-20 07:04:52 +0000 UTC" firstStartedPulling="2026-02-20 07:04:53.720957485 +0000 UTC m=+1450.492246463" lastFinishedPulling="2026-02-20 07:04:54.252033482 +0000 UTC m=+1451.023322460" observedRunningTime="2026-02-20 07:04:54.880086189 +0000 UTC m=+1451.651375167" watchObservedRunningTime="2026-02-20 07:04:54.885770607 +0000 UTC m=+1451.657059584" Feb 20 07:05:04 crc kubenswrapper[4492]: I0220 07:05:04.434224 4492 scope.go:117] "RemoveContainer" containerID="ec29179caa35b44ea7c3fd4cefbb1301d7d2414919c86493041d341ae008746c" Feb 20 07:05:04 crc kubenswrapper[4492]: I0220 07:05:04.465748 4492 scope.go:117] "RemoveContainer" containerID="9872e866b203c8b65f2f4650d664f7d3d7e29ec00a9f07461b81a2f5e3a19adb" Feb 20 07:05:04 crc kubenswrapper[4492]: I0220 07:05:04.491759 4492 scope.go:117] "RemoveContainer" containerID="b8bcadccc077a072ddf622973e188bafb6c91e58b1586239ec90786d7b7cad7a" Feb 20 07:05:04 crc kubenswrapper[4492]: I0220 07:05:04.514148 4492 scope.go:117] "RemoveContainer" containerID="50b83c0ff9608ee76e0d6bfbb33906d4cb0567180049e7d00ffc2c3d6acc95d6" Feb 20 07:05:09 crc kubenswrapper[4492]: I0220 07:05:09.311769 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:05:09 crc kubenswrapper[4492]: I0220 07:05:09.312564 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.059467 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c239-account-create-update-kfsgt"] Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.064914 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-0ad5-account-create-update-7lnn9"] Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.070035 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-qxtqh"] Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.075115 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g4ndh"] Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.081408 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-qxtqh"] Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.086266 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-0ad5-account-create-update-7lnn9"] Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.091099 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c239-account-create-update-kfsgt"] Feb 20 07:05:12 crc kubenswrapper[4492]: I0220 07:05:12.095747 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g4ndh"] Feb 20 07:05:13 crc kubenswrapper[4492]: I0220 07:05:13.570188 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d794150-2156-4a32-8f44-4fe9a4b45a88" path="/var/lib/kubelet/pods/0d794150-2156-4a32-8f44-4fe9a4b45a88/volumes" Feb 20 07:05:13 crc kubenswrapper[4492]: I0220 07:05:13.572799 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e25e689-57da-4dc7-a3e6-1aabddb8acce" path="/var/lib/kubelet/pods/3e25e689-57da-4dc7-a3e6-1aabddb8acce/volumes" Feb 20 07:05:13 crc kubenswrapper[4492]: I0220 07:05:13.574210 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4be5bc11-b1d3-4456-b3a7-060de7a1bc2f" path="/var/lib/kubelet/pods/4be5bc11-b1d3-4456-b3a7-060de7a1bc2f/volumes" Feb 20 07:05:13 crc kubenswrapper[4492]: I0220 07:05:13.575515 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6513562e-a81e-4d6f-9ba9-1d7453aec92c" path="/var/lib/kubelet/pods/6513562e-a81e-4d6f-9ba9-1d7453aec92c/volumes" Feb 20 07:05:16 crc kubenswrapper[4492]: I0220 07:05:16.026846 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6d7c-account-create-update-p6txs"] Feb 20 07:05:16 crc kubenswrapper[4492]: I0220 07:05:16.035522 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kcmd6"] Feb 20 07:05:16 crc kubenswrapper[4492]: I0220 07:05:16.043049 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6d7c-account-create-update-p6txs"] Feb 20 07:05:16 crc kubenswrapper[4492]: I0220 07:05:16.050825 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kcmd6"] Feb 20 07:05:17 crc kubenswrapper[4492]: I0220 07:05:17.569650 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11551b3b-e2e7-4c28-8886-b4bfe872010e" path="/var/lib/kubelet/pods/11551b3b-e2e7-4c28-8886-b4bfe872010e/volumes" Feb 20 07:05:17 crc kubenswrapper[4492]: I0220 07:05:17.571935 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc" path="/var/lib/kubelet/pods/d2ef36d8-1d3d-4bad-ad8b-f8aeb2fadccc/volumes" Feb 20 07:05:18 crc kubenswrapper[4492]: I0220 07:05:18.062466 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-hhsgt"] Feb 20 07:05:18 crc kubenswrapper[4492]: I0220 07:05:18.068189 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-hhsgt"] Feb 20 07:05:19 crc kubenswrapper[4492]: I0220 07:05:19.568305 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e691f76-c662-4efa-a5e8-715c14abe03d" path="/var/lib/kubelet/pods/9e691f76-c662-4efa-a5e8-715c14abe03d/volumes" Feb 20 07:05:39 crc kubenswrapper[4492]: I0220 07:05:39.311201 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:05:39 crc kubenswrapper[4492]: I0220 07:05:39.311933 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.069777 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-qjqbf"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.080401 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-91fd-account-create-update-68gbg"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.095788 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-02b5-account-create-update-nlqzc"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.101842 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-02b5-account-create-update-nlqzc"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.109752 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-91fd-account-create-update-68gbg"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.119565 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-q26vt"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.125775 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-qjqbf"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.131151 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-5znh9"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.168459 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-5znh9"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.180738 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-q26vt"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.188965 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-c12a-account-create-update-qx47p"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.195905 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-c12a-account-create-update-qx47p"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.203667 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-lq7fc"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.217668 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7e35-account-create-update-22fl8"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.229308 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-lq7fc"] Feb 20 07:05:40 crc kubenswrapper[4492]: I0220 07:05:40.239697 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7e35-account-create-update-22fl8"] Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.579185 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aba8976-47dd-4f37-9e4d-e561b7f05495" path="/var/lib/kubelet/pods/0aba8976-47dd-4f37-9e4d-e561b7f05495/volumes" Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.580644 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21c3e515-5917-4360-8ab2-5fc920f8f537" path="/var/lib/kubelet/pods/21c3e515-5917-4360-8ab2-5fc920f8f537/volumes" Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.583066 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d8c5715-e9fb-4eef-994f-04aeb6b18864" path="/var/lib/kubelet/pods/2d8c5715-e9fb-4eef-994f-04aeb6b18864/volumes" Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.584434 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c067dc9-f4a3-46d3-bb5f-bc8f36312a79" path="/var/lib/kubelet/pods/5c067dc9-f4a3-46d3-bb5f-bc8f36312a79/volumes" Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.586351 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64054bf5-de18-49b3-93c5-e5081ef122f8" path="/var/lib/kubelet/pods/64054bf5-de18-49b3-93c5-e5081ef122f8/volumes" Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.587790 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84053541-a741-4dd3-8b4c-420cfa5da896" path="/var/lib/kubelet/pods/84053541-a741-4dd3-8b4c-420cfa5da896/volumes" Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.590575 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e8d47d-5fa9-4cae-8164-8b5bf089c82f" path="/var/lib/kubelet/pods/b1e8d47d-5fa9-4cae-8164-8b5bf089c82f/volumes" Feb 20 07:05:41 crc kubenswrapper[4492]: I0220 07:05:41.591507 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d59d1326-3f4e-4429-93ee-4b983de7ff82" path="/var/lib/kubelet/pods/d59d1326-3f4e-4429-93ee-4b983de7ff82/volumes" Feb 20 07:06:01 crc kubenswrapper[4492]: I0220 07:06:01.041812 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-scs24"] Feb 20 07:06:01 crc kubenswrapper[4492]: I0220 07:06:01.048880 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-scs24"] Feb 20 07:06:01 crc kubenswrapper[4492]: I0220 07:06:01.584190 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e4ff2bf-0f92-4271-bc11-e07604d93c5b" path="/var/lib/kubelet/pods/8e4ff2bf-0f92-4271-bc11-e07604d93c5b/volumes" Feb 20 07:06:03 crc kubenswrapper[4492]: I0220 07:06:03.863906 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fpbgr"] Feb 20 07:06:03 crc kubenswrapper[4492]: I0220 07:06:03.867317 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:03 crc kubenswrapper[4492]: I0220 07:06:03.875865 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fpbgr"] Feb 20 07:06:03 crc kubenswrapper[4492]: I0220 07:06:03.925207 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-utilities\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:03 crc kubenswrapper[4492]: I0220 07:06:03.925330 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-catalog-content\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:03 crc kubenswrapper[4492]: I0220 07:06:03.925401 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2hcr\" (UniqueName: \"kubernetes.io/projected/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-kube-api-access-l2hcr\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.027545 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2hcr\" (UniqueName: \"kubernetes.io/projected/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-kube-api-access-l2hcr\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.027670 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-utilities\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.027825 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-catalog-content\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.028231 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-utilities\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.028316 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-catalog-content\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.047582 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2hcr\" (UniqueName: \"kubernetes.io/projected/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-kube-api-access-l2hcr\") pod \"certified-operators-fpbgr\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.183051 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.579437 4492 scope.go:117] "RemoveContainer" containerID="21fa4b0532cfd3e0efd8c7c742e61898fed119901fe4c38113a0db67b6d6c1be" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.612694 4492 scope.go:117] "RemoveContainer" containerID="d1212b8c0ed3fc190301b7d2e44b1802788c6ef622328e6c8751ffbf948f4305" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.619104 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fpbgr"] Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.691442 4492 scope.go:117] "RemoveContainer" containerID="0dfc80c67ab2eeb6dac4d419d3ee27c2b874ccef518b98f64045ea6b4346bb5c" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.715756 4492 scope.go:117] "RemoveContainer" containerID="e8c20a4208f06b525332e7679b73ca42d6f250ba8c9fab55097f421ae3dddc0c" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.734666 4492 scope.go:117] "RemoveContainer" containerID="db062a294b75d86880b4bdaba992eacadeecd70bf9478ecd6a4929b0c7a63593" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.765089 4492 scope.go:117] "RemoveContainer" containerID="60fee276a052c1f522f17721904a700120bf9ac76abc5f17f1a4ca7c3d821c35" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.810663 4492 scope.go:117] "RemoveContainer" containerID="5e4128e68b709854f0df792e0dd075a47289f3b923a4db57f5da0f575394dfc5" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.830000 4492 scope.go:117] "RemoveContainer" containerID="e1d88bafc59d2a7026a983fdf0fbaf0d489bae206d1f432191e15b7c691d0d26" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.864446 4492 scope.go:117] "RemoveContainer" containerID="5b1471e28cb25ff9e5cbe97b972ca8a3e60566099b74dbd28911c8e42fa1d0f0" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.885183 4492 scope.go:117] "RemoveContainer" containerID="2a997b0e91405df93db58d3c407de0407413b6972da3636b4babc2909f92d11f" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.905027 4492 scope.go:117] "RemoveContainer" containerID="84c58b2555461bb9a56fd038c6bbf8ac21ea32306af96e0609692552c01e09f0" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.932951 4492 scope.go:117] "RemoveContainer" containerID="f29efd2d96984f6830b5997d3c665cd0a882f0d7956243176c9841b763c39f6c" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.961930 4492 scope.go:117] "RemoveContainer" containerID="8b98e9676fcb838bc3cccc1cd17480c8b57e9f708a1b95d7ca19ee900fe8155a" Feb 20 07:06:04 crc kubenswrapper[4492]: I0220 07:06:04.982932 4492 scope.go:117] "RemoveContainer" containerID="3ce12d434ad1ecd2f28b179736935ad1433b0d8a589f939bd4357a21ff1e4f6a" Feb 20 07:06:05 crc kubenswrapper[4492]: I0220 07:06:05.026447 4492 scope.go:117] "RemoveContainer" containerID="aa7ec4a6bf75470292e87ceb15bb61eca26af293374afb3c45bdaad1cba963a6" Feb 20 07:06:05 crc kubenswrapper[4492]: I0220 07:06:05.047174 4492 scope.go:117] "RemoveContainer" containerID="88ff3b8feb37a1ce94ecb080e21bdfa3c389947bbd6a378c8d7617bffa8da2ed" Feb 20 07:06:05 crc kubenswrapper[4492]: I0220 07:06:05.117938 4492 scope.go:117] "RemoveContainer" containerID="e1e5e6202df7c982b25cf3e0d1b04f0cfe3aab232c603f22b823ef1e94cf01cd" Feb 20 07:06:05 crc kubenswrapper[4492]: I0220 07:06:05.160139 4492 scope.go:117] "RemoveContainer" containerID="33184a9d99961a80661fe2f9c2a097634967d28b467cc60def086e864a6e75bf" Feb 20 07:06:05 crc kubenswrapper[4492]: I0220 07:06:05.585356 4492 generic.go:334] "Generic (PLEG): container finished" podID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerID="d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4" exitCode=0 Feb 20 07:06:05 crc kubenswrapper[4492]: I0220 07:06:05.585409 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpbgr" event={"ID":"055ba5d4-eee0-449c-82b0-0ef52c8eb94a","Type":"ContainerDied","Data":"d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4"} Feb 20 07:06:05 crc kubenswrapper[4492]: I0220 07:06:05.585437 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpbgr" event={"ID":"055ba5d4-eee0-449c-82b0-0ef52c8eb94a","Type":"ContainerStarted","Data":"302c261510789b2b1f275825e9c0f1be866c0ff5e2cf10f248c27a95ef78fae3"} Feb 20 07:06:06 crc kubenswrapper[4492]: I0220 07:06:06.599333 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpbgr" event={"ID":"055ba5d4-eee0-449c-82b0-0ef52c8eb94a","Type":"ContainerStarted","Data":"538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a"} Feb 20 07:06:07 crc kubenswrapper[4492]: I0220 07:06:07.609672 4492 generic.go:334] "Generic (PLEG): container finished" podID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerID="538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a" exitCode=0 Feb 20 07:06:07 crc kubenswrapper[4492]: I0220 07:06:07.609717 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpbgr" event={"ID":"055ba5d4-eee0-449c-82b0-0ef52c8eb94a","Type":"ContainerDied","Data":"538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a"} Feb 20 07:06:08 crc kubenswrapper[4492]: I0220 07:06:08.625069 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpbgr" event={"ID":"055ba5d4-eee0-449c-82b0-0ef52c8eb94a","Type":"ContainerStarted","Data":"c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d"} Feb 20 07:06:08 crc kubenswrapper[4492]: I0220 07:06:08.644136 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fpbgr" podStartSLOduration=3.101213785 podStartE2EDuration="5.64412017s" podCreationTimestamp="2026-02-20 07:06:03 +0000 UTC" firstStartedPulling="2026-02-20 07:06:05.587699174 +0000 UTC m=+1522.358988152" lastFinishedPulling="2026-02-20 07:06:08.130605558 +0000 UTC m=+1524.901894537" observedRunningTime="2026-02-20 07:06:08.641938397 +0000 UTC m=+1525.413227375" watchObservedRunningTime="2026-02-20 07:06:08.64412017 +0000 UTC m=+1525.415409149" Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.311303 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.311371 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.311424 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.312323 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.312383 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" gracePeriod=600 Feb 20 07:06:09 crc kubenswrapper[4492]: E0220 07:06:09.444178 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.635353 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" exitCode=0 Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.635389 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe"} Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.635449 4492 scope.go:117] "RemoveContainer" containerID="ab8284c72f81bddf931f4abee5f04c18596ed9e487d4c6a5d3e615ec918dcbba" Feb 20 07:06:09 crc kubenswrapper[4492]: I0220 07:06:09.636504 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:06:09 crc kubenswrapper[4492]: E0220 07:06:09.637208 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:06:11 crc kubenswrapper[4492]: I0220 07:06:11.027003 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-x5r2g"] Feb 20 07:06:11 crc kubenswrapper[4492]: I0220 07:06:11.033941 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-x5r2g"] Feb 20 07:06:11 crc kubenswrapper[4492]: I0220 07:06:11.566943 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaa400f5-22b0-44f7-9655-f8b3bd1f8467" path="/var/lib/kubelet/pods/eaa400f5-22b0-44f7-9655-f8b3bd1f8467/volumes" Feb 20 07:06:14 crc kubenswrapper[4492]: I0220 07:06:14.183869 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:14 crc kubenswrapper[4492]: I0220 07:06:14.184165 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:14 crc kubenswrapper[4492]: I0220 07:06:14.222724 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:14 crc kubenswrapper[4492]: I0220 07:06:14.730777 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:14 crc kubenswrapper[4492]: I0220 07:06:14.769910 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fpbgr"] Feb 20 07:06:16 crc kubenswrapper[4492]: I0220 07:06:16.712157 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fpbgr" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="registry-server" containerID="cri-o://c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d" gracePeriod=2 Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.114341 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.302929 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-utilities\") pod \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.303283 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2hcr\" (UniqueName: \"kubernetes.io/projected/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-kube-api-access-l2hcr\") pod \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.303512 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-catalog-content\") pod \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\" (UID: \"055ba5d4-eee0-449c-82b0-0ef52c8eb94a\") " Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.303745 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-utilities" (OuterVolumeSpecName: "utilities") pod "055ba5d4-eee0-449c-82b0-0ef52c8eb94a" (UID: "055ba5d4-eee0-449c-82b0-0ef52c8eb94a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.304802 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.318189 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-kube-api-access-l2hcr" (OuterVolumeSpecName: "kube-api-access-l2hcr") pod "055ba5d4-eee0-449c-82b0-0ef52c8eb94a" (UID: "055ba5d4-eee0-449c-82b0-0ef52c8eb94a"). InnerVolumeSpecName "kube-api-access-l2hcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.346617 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "055ba5d4-eee0-449c-82b0-0ef52c8eb94a" (UID: "055ba5d4-eee0-449c-82b0-0ef52c8eb94a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.407329 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2hcr\" (UniqueName: \"kubernetes.io/projected/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-kube-api-access-l2hcr\") on node \"crc\" DevicePath \"\"" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.407356 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/055ba5d4-eee0-449c-82b0-0ef52c8eb94a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.722342 4492 generic.go:334] "Generic (PLEG): container finished" podID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerID="c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d" exitCode=0 Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.722428 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fpbgr" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.722419 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpbgr" event={"ID":"055ba5d4-eee0-449c-82b0-0ef52c8eb94a","Type":"ContainerDied","Data":"c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d"} Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.722741 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fpbgr" event={"ID":"055ba5d4-eee0-449c-82b0-0ef52c8eb94a","Type":"ContainerDied","Data":"302c261510789b2b1f275825e9c0f1be866c0ff5e2cf10f248c27a95ef78fae3"} Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.722764 4492 scope.go:117] "RemoveContainer" containerID="c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.746891 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fpbgr"] Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.752710 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fpbgr"] Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.756279 4492 scope.go:117] "RemoveContainer" containerID="538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.777386 4492 scope.go:117] "RemoveContainer" containerID="d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.805415 4492 scope.go:117] "RemoveContainer" containerID="c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d" Feb 20 07:06:17 crc kubenswrapper[4492]: E0220 07:06:17.805867 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d\": container with ID starting with c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d not found: ID does not exist" containerID="c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.805964 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d"} err="failed to get container status \"c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d\": rpc error: code = NotFound desc = could not find container \"c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d\": container with ID starting with c24dc7b7ad4f051a6804b8e8f9e3d39389637125a1a16b19aca0cbabec5a717d not found: ID does not exist" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.806045 4492 scope.go:117] "RemoveContainer" containerID="538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a" Feb 20 07:06:17 crc kubenswrapper[4492]: E0220 07:06:17.806354 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a\": container with ID starting with 538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a not found: ID does not exist" containerID="538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.806382 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a"} err="failed to get container status \"538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a\": rpc error: code = NotFound desc = could not find container \"538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a\": container with ID starting with 538681cfdefa83c3dd15544d129e1ef69bf36ec70c14384eac7a74295c61947a not found: ID does not exist" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.806398 4492 scope.go:117] "RemoveContainer" containerID="d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4" Feb 20 07:06:17 crc kubenswrapper[4492]: E0220 07:06:17.806665 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4\": container with ID starting with d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4 not found: ID does not exist" containerID="d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4" Feb 20 07:06:17 crc kubenswrapper[4492]: I0220 07:06:17.806709 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4"} err="failed to get container status \"d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4\": rpc error: code = NotFound desc = could not find container \"d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4\": container with ID starting with d5ce4d8ca58f70ed0abd1a8a0a1481ae184c5b37f34fe5221c90eba618fb44a4 not found: ID does not exist" Feb 20 07:06:19 crc kubenswrapper[4492]: I0220 07:06:19.566949 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" path="/var/lib/kubelet/pods/055ba5d4-eee0-449c-82b0-0ef52c8eb94a/volumes" Feb 20 07:06:20 crc kubenswrapper[4492]: I0220 07:06:20.556739 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:06:20 crc kubenswrapper[4492]: E0220 07:06:20.557048 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:06:35 crc kubenswrapper[4492]: I0220 07:06:35.557380 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:06:35 crc kubenswrapper[4492]: E0220 07:06:35.558223 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.652190 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2v45m"] Feb 20 07:06:39 crc kubenswrapper[4492]: E0220 07:06:39.653448 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="extract-content" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.653488 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="extract-content" Feb 20 07:06:39 crc kubenswrapper[4492]: E0220 07:06:39.653517 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="registry-server" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.653524 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="registry-server" Feb 20 07:06:39 crc kubenswrapper[4492]: E0220 07:06:39.653534 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="extract-utilities" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.653542 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="extract-utilities" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.653815 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="055ba5d4-eee0-449c-82b0-0ef52c8eb94a" containerName="registry-server" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.655765 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.665315 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2v45m"] Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.780893 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6c7m\" (UniqueName: \"kubernetes.io/projected/65910b40-8daf-4cbd-91ca-e5453ec08bba-kube-api-access-q6c7m\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.780990 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-catalog-content\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.781015 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-utilities\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.882132 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-catalog-content\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.882169 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-utilities\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.882274 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6c7m\" (UniqueName: \"kubernetes.io/projected/65910b40-8daf-4cbd-91ca-e5453ec08bba-kube-api-access-q6c7m\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.882760 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-catalog-content\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.882808 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-utilities\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.908604 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6c7m\" (UniqueName: \"kubernetes.io/projected/65910b40-8daf-4cbd-91ca-e5453ec08bba-kube-api-access-q6c7m\") pod \"community-operators-2v45m\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:39 crc kubenswrapper[4492]: I0220 07:06:39.971283 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:40 crc kubenswrapper[4492]: W0220 07:06:40.411705 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65910b40_8daf_4cbd_91ca_e5453ec08bba.slice/crio-7406ce61aec7d3bec5197c8e955d87654f7c39dbab1766fb4599e62a55197339 WatchSource:0}: Error finding container 7406ce61aec7d3bec5197c8e955d87654f7c39dbab1766fb4599e62a55197339: Status 404 returned error can't find the container with id 7406ce61aec7d3bec5197c8e955d87654f7c39dbab1766fb4599e62a55197339 Feb 20 07:06:40 crc kubenswrapper[4492]: I0220 07:06:40.420098 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2v45m"] Feb 20 07:06:40 crc kubenswrapper[4492]: I0220 07:06:40.945974 4492 generic.go:334] "Generic (PLEG): container finished" podID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerID="747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947" exitCode=0 Feb 20 07:06:40 crc kubenswrapper[4492]: I0220 07:06:40.946083 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2v45m" event={"ID":"65910b40-8daf-4cbd-91ca-e5453ec08bba","Type":"ContainerDied","Data":"747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947"} Feb 20 07:06:40 crc kubenswrapper[4492]: I0220 07:06:40.946405 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2v45m" event={"ID":"65910b40-8daf-4cbd-91ca-e5453ec08bba","Type":"ContainerStarted","Data":"7406ce61aec7d3bec5197c8e955d87654f7c39dbab1766fb4599e62a55197339"} Feb 20 07:06:41 crc kubenswrapper[4492]: I0220 07:06:41.958857 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2v45m" event={"ID":"65910b40-8daf-4cbd-91ca-e5453ec08bba","Type":"ContainerStarted","Data":"da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8"} Feb 20 07:06:42 crc kubenswrapper[4492]: I0220 07:06:42.038861 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2s268"] Feb 20 07:06:42 crc kubenswrapper[4492]: I0220 07:06:42.045360 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2s268"] Feb 20 07:06:42 crc kubenswrapper[4492]: I0220 07:06:42.968932 4492 generic.go:334] "Generic (PLEG): container finished" podID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerID="da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8" exitCode=0 Feb 20 07:06:42 crc kubenswrapper[4492]: I0220 07:06:42.969017 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2v45m" event={"ID":"65910b40-8daf-4cbd-91ca-e5453ec08bba","Type":"ContainerDied","Data":"da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8"} Feb 20 07:06:43 crc kubenswrapper[4492]: I0220 07:06:43.571067 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d697f3d-7b65-4983-8602-de839d5caa82" path="/var/lib/kubelet/pods/4d697f3d-7b65-4983-8602-de839d5caa82/volumes" Feb 20 07:06:43 crc kubenswrapper[4492]: I0220 07:06:43.981518 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2v45m" event={"ID":"65910b40-8daf-4cbd-91ca-e5453ec08bba","Type":"ContainerStarted","Data":"1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42"} Feb 20 07:06:43 crc kubenswrapper[4492]: I0220 07:06:43.997041 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2v45m" podStartSLOduration=2.471095781 podStartE2EDuration="4.997020413s" podCreationTimestamp="2026-02-20 07:06:39 +0000 UTC" firstStartedPulling="2026-02-20 07:06:40.947938101 +0000 UTC m=+1557.719227079" lastFinishedPulling="2026-02-20 07:06:43.473862732 +0000 UTC m=+1560.245151711" observedRunningTime="2026-02-20 07:06:43.995639178 +0000 UTC m=+1560.766928156" watchObservedRunningTime="2026-02-20 07:06:43.997020413 +0000 UTC m=+1560.768309391" Feb 20 07:06:46 crc kubenswrapper[4492]: I0220 07:06:46.557524 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:06:46 crc kubenswrapper[4492]: E0220 07:06:46.559050 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.037800 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wb525"] Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.040333 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.048150 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6qxl\" (UniqueName: \"kubernetes.io/projected/73fc93b1-fb18-4271-af6f-ad1278738fdc-kube-api-access-w6qxl\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.048229 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-utilities\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.048300 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-catalog-content\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.049208 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb525"] Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.150182 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6qxl\" (UniqueName: \"kubernetes.io/projected/73fc93b1-fb18-4271-af6f-ad1278738fdc-kube-api-access-w6qxl\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.150267 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-utilities\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.150325 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-catalog-content\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.150857 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-catalog-content\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.151027 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-utilities\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.169560 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6qxl\" (UniqueName: \"kubernetes.io/projected/73fc93b1-fb18-4271-af6f-ad1278738fdc-kube-api-access-w6qxl\") pod \"redhat-marketplace-wb525\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.365488 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:47 crc kubenswrapper[4492]: I0220 07:06:47.814254 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb525"] Feb 20 07:06:48 crc kubenswrapper[4492]: I0220 07:06:48.039215 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerStarted","Data":"d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0"} Feb 20 07:06:48 crc kubenswrapper[4492]: I0220 07:06:48.039268 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerStarted","Data":"171398cf4539120d7184cae366ee0857ee3d29683344d2cd5fb14cd728376077"} Feb 20 07:06:49 crc kubenswrapper[4492]: I0220 07:06:49.050058 4492 generic.go:334] "Generic (PLEG): container finished" podID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerID="d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0" exitCode=0 Feb 20 07:06:49 crc kubenswrapper[4492]: I0220 07:06:49.050116 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerDied","Data":"d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0"} Feb 20 07:06:49 crc kubenswrapper[4492]: I0220 07:06:49.971952 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:49 crc kubenswrapper[4492]: I0220 07:06:49.972335 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:50 crc kubenswrapper[4492]: I0220 07:06:50.012831 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:50 crc kubenswrapper[4492]: I0220 07:06:50.066541 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerStarted","Data":"e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b"} Feb 20 07:06:50 crc kubenswrapper[4492]: I0220 07:06:50.129778 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:51 crc kubenswrapper[4492]: I0220 07:06:51.074434 4492 generic.go:334] "Generic (PLEG): container finished" podID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerID="e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b" exitCode=0 Feb 20 07:06:51 crc kubenswrapper[4492]: I0220 07:06:51.074487 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerDied","Data":"e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b"} Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.087650 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerStarted","Data":"99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b"} Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.102196 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wb525" podStartSLOduration=2.582981807 podStartE2EDuration="5.102180268s" podCreationTimestamp="2026-02-20 07:06:47 +0000 UTC" firstStartedPulling="2026-02-20 07:06:49.051613227 +0000 UTC m=+1565.822902205" lastFinishedPulling="2026-02-20 07:06:51.570811688 +0000 UTC m=+1568.342100666" observedRunningTime="2026-02-20 07:06:52.100227547 +0000 UTC m=+1568.871516524" watchObservedRunningTime="2026-02-20 07:06:52.102180268 +0000 UTC m=+1568.873469246" Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.429017 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2v45m"] Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.429207 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2v45m" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="registry-server" containerID="cri-o://1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42" gracePeriod=2 Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.855373 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.968319 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-utilities\") pod \"65910b40-8daf-4cbd-91ca-e5453ec08bba\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.968586 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-catalog-content\") pod \"65910b40-8daf-4cbd-91ca-e5453ec08bba\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.968765 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6c7m\" (UniqueName: \"kubernetes.io/projected/65910b40-8daf-4cbd-91ca-e5453ec08bba-kube-api-access-q6c7m\") pod \"65910b40-8daf-4cbd-91ca-e5453ec08bba\" (UID: \"65910b40-8daf-4cbd-91ca-e5453ec08bba\") " Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.969075 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-utilities" (OuterVolumeSpecName: "utilities") pod "65910b40-8daf-4cbd-91ca-e5453ec08bba" (UID: "65910b40-8daf-4cbd-91ca-e5453ec08bba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.969355 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:06:52 crc kubenswrapper[4492]: I0220 07:06:52.982662 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65910b40-8daf-4cbd-91ca-e5453ec08bba-kube-api-access-q6c7m" (OuterVolumeSpecName: "kube-api-access-q6c7m") pod "65910b40-8daf-4cbd-91ca-e5453ec08bba" (UID: "65910b40-8daf-4cbd-91ca-e5453ec08bba"). InnerVolumeSpecName "kube-api-access-q6c7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.022058 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65910b40-8daf-4cbd-91ca-e5453ec08bba" (UID: "65910b40-8daf-4cbd-91ca-e5453ec08bba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.072083 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65910b40-8daf-4cbd-91ca-e5453ec08bba-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.072117 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6c7m\" (UniqueName: \"kubernetes.io/projected/65910b40-8daf-4cbd-91ca-e5453ec08bba-kube-api-access-q6c7m\") on node \"crc\" DevicePath \"\"" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.103279 4492 generic.go:334] "Generic (PLEG): container finished" podID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerID="1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42" exitCode=0 Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.103349 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2v45m" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.103372 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2v45m" event={"ID":"65910b40-8daf-4cbd-91ca-e5453ec08bba","Type":"ContainerDied","Data":"1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42"} Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.103831 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2v45m" event={"ID":"65910b40-8daf-4cbd-91ca-e5453ec08bba","Type":"ContainerDied","Data":"7406ce61aec7d3bec5197c8e955d87654f7c39dbab1766fb4599e62a55197339"} Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.103871 4492 scope.go:117] "RemoveContainer" containerID="1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.130612 4492 scope.go:117] "RemoveContainer" containerID="da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.134555 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2v45m"] Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.144595 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2v45m"] Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.149656 4492 scope.go:117] "RemoveContainer" containerID="747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.185744 4492 scope.go:117] "RemoveContainer" containerID="1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42" Feb 20 07:06:53 crc kubenswrapper[4492]: E0220 07:06:53.186086 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42\": container with ID starting with 1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42 not found: ID does not exist" containerID="1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.186119 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42"} err="failed to get container status \"1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42\": rpc error: code = NotFound desc = could not find container \"1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42\": container with ID starting with 1aa8d703e22b7a91184456b496d2770f9e72490688ae0595be4be2687a427b42 not found: ID does not exist" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.186143 4492 scope.go:117] "RemoveContainer" containerID="da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8" Feb 20 07:06:53 crc kubenswrapper[4492]: E0220 07:06:53.186413 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8\": container with ID starting with da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8 not found: ID does not exist" containerID="da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.186438 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8"} err="failed to get container status \"da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8\": rpc error: code = NotFound desc = could not find container \"da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8\": container with ID starting with da8b342aab1e5df9f6a9d1f76f221acaa5f8fdb5095959b4b7b8a04015414aa8 not found: ID does not exist" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.186452 4492 scope.go:117] "RemoveContainer" containerID="747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947" Feb 20 07:06:53 crc kubenswrapper[4492]: E0220 07:06:53.186694 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947\": container with ID starting with 747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947 not found: ID does not exist" containerID="747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.186713 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947"} err="failed to get container status \"747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947\": rpc error: code = NotFound desc = could not find container \"747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947\": container with ID starting with 747bf77207461326a260ce078aad760dbb974d0ef8f4317b1ae05a7216070947 not found: ID does not exist" Feb 20 07:06:53 crc kubenswrapper[4492]: I0220 07:06:53.568374 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" path="/var/lib/kubelet/pods/65910b40-8daf-4cbd-91ca-e5453ec08bba/volumes" Feb 20 07:06:57 crc kubenswrapper[4492]: I0220 07:06:57.366612 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:57 crc kubenswrapper[4492]: I0220 07:06:57.367240 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:57 crc kubenswrapper[4492]: I0220 07:06:57.408001 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:58 crc kubenswrapper[4492]: I0220 07:06:58.188312 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:06:58 crc kubenswrapper[4492]: I0220 07:06:58.425969 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb525"] Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.163327 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wb525" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="registry-server" containerID="cri-o://99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b" gracePeriod=2 Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.535464 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.556889 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:07:00 crc kubenswrapper[4492]: E0220 07:07:00.557191 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.642029 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6qxl\" (UniqueName: \"kubernetes.io/projected/73fc93b1-fb18-4271-af6f-ad1278738fdc-kube-api-access-w6qxl\") pod \"73fc93b1-fb18-4271-af6f-ad1278738fdc\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.642236 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-utilities\") pod \"73fc93b1-fb18-4271-af6f-ad1278738fdc\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.642289 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-catalog-content\") pod \"73fc93b1-fb18-4271-af6f-ad1278738fdc\" (UID: \"73fc93b1-fb18-4271-af6f-ad1278738fdc\") " Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.642971 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-utilities" (OuterVolumeSpecName: "utilities") pod "73fc93b1-fb18-4271-af6f-ad1278738fdc" (UID: "73fc93b1-fb18-4271-af6f-ad1278738fdc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.658856 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73fc93b1-fb18-4271-af6f-ad1278738fdc-kube-api-access-w6qxl" (OuterVolumeSpecName: "kube-api-access-w6qxl") pod "73fc93b1-fb18-4271-af6f-ad1278738fdc" (UID: "73fc93b1-fb18-4271-af6f-ad1278738fdc"). InnerVolumeSpecName "kube-api-access-w6qxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.662905 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73fc93b1-fb18-4271-af6f-ad1278738fdc" (UID: "73fc93b1-fb18-4271-af6f-ad1278738fdc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.745373 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6qxl\" (UniqueName: \"kubernetes.io/projected/73fc93b1-fb18-4271-af6f-ad1278738fdc-kube-api-access-w6qxl\") on node \"crc\" DevicePath \"\"" Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.745413 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:07:00 crc kubenswrapper[4492]: I0220 07:07:00.745428 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73fc93b1-fb18-4271-af6f-ad1278738fdc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.174737 4492 generic.go:334] "Generic (PLEG): container finished" podID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerID="99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b" exitCode=0 Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.174795 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerDied","Data":"99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b"} Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.174841 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wb525" event={"ID":"73fc93b1-fb18-4271-af6f-ad1278738fdc","Type":"ContainerDied","Data":"171398cf4539120d7184cae366ee0857ee3d29683344d2cd5fb14cd728376077"} Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.174861 4492 scope.go:117] "RemoveContainer" containerID="99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.176195 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wb525" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.194325 4492 scope.go:117] "RemoveContainer" containerID="e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.214740 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb525"] Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.224461 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wb525"] Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.228273 4492 scope.go:117] "RemoveContainer" containerID="d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.258129 4492 scope.go:117] "RemoveContainer" containerID="99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b" Feb 20 07:07:01 crc kubenswrapper[4492]: E0220 07:07:01.258586 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b\": container with ID starting with 99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b not found: ID does not exist" containerID="99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.258635 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b"} err="failed to get container status \"99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b\": rpc error: code = NotFound desc = could not find container \"99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b\": container with ID starting with 99903a4ac9d1a19ba8ac60d66ce14074e5b9bc4989b097066538dcb13cea1c5b not found: ID does not exist" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.258668 4492 scope.go:117] "RemoveContainer" containerID="e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b" Feb 20 07:07:01 crc kubenswrapper[4492]: E0220 07:07:01.259116 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b\": container with ID starting with e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b not found: ID does not exist" containerID="e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.259145 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b"} err="failed to get container status \"e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b\": rpc error: code = NotFound desc = could not find container \"e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b\": container with ID starting with e0ac4266f625fa2e5e22d382dc309ef26374d277a3516eb7f3bfd37abba2c18b not found: ID does not exist" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.259192 4492 scope.go:117] "RemoveContainer" containerID="d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0" Feb 20 07:07:01 crc kubenswrapper[4492]: E0220 07:07:01.259577 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0\": container with ID starting with d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0 not found: ID does not exist" containerID="d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.259690 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0"} err="failed to get container status \"d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0\": rpc error: code = NotFound desc = could not find container \"d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0\": container with ID starting with d69218f3ac96f93e7c7c8f18443e6514a76362a4252663f5b31db8cb193466d0 not found: ID does not exist" Feb 20 07:07:01 crc kubenswrapper[4492]: I0220 07:07:01.566737 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" path="/var/lib/kubelet/pods/73fc93b1-fb18-4271-af6f-ad1278738fdc/volumes" Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.071222 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-s246x"] Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.079331 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gjtqr"] Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.086512 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-s246x"] Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.092319 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gjtqr"] Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.098259 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-blbr6"] Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.103243 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-blbr6"] Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.438862 4492 scope.go:117] "RemoveContainer" containerID="d40bf6360c8b922955e98ea57d1c630a3d16f148de5c5d4a4e824041c59ebf58" Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.461373 4492 scope.go:117] "RemoveContainer" containerID="7783b346b02ae90a0e391645933bc3abef3edc4f1a4e94804ba624f030168e71" Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.567992 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b719757-21ff-4967-a478-287c671d2bc7" path="/var/lib/kubelet/pods/5b719757-21ff-4967-a478-287c671d2bc7/volumes" Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.569507 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ee3d5af-cee2-4398-b8c8-969d7aa69fdc" path="/var/lib/kubelet/pods/7ee3d5af-cee2-4398-b8c8-969d7aa69fdc/volumes" Feb 20 07:07:05 crc kubenswrapper[4492]: I0220 07:07:05.571913 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a83f66a1-3f7a-4ff0-bb95-3eff19f92019" path="/var/lib/kubelet/pods/a83f66a1-3f7a-4ff0-bb95-3eff19f92019/volumes" Feb 20 07:07:06 crc kubenswrapper[4492]: I0220 07:07:06.220246 4492 generic.go:334] "Generic (PLEG): container finished" podID="d0c74792-a006-4752-8403-70947f6fec63" containerID="9847b6eca3297f0d076cfef8bd706cd06a64f259904ab9fb7f1f54bbfaeb4db3" exitCode=0 Feb 20 07:07:06 crc kubenswrapper[4492]: I0220 07:07:06.220335 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" event={"ID":"d0c74792-a006-4752-8403-70947f6fec63","Type":"ContainerDied","Data":"9847b6eca3297f0d076cfef8bd706cd06a64f259904ab9fb7f1f54bbfaeb4db3"} Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.630335 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.808656 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grgn8\" (UniqueName: \"kubernetes.io/projected/d0c74792-a006-4752-8403-70947f6fec63-kube-api-access-grgn8\") pod \"d0c74792-a006-4752-8403-70947f6fec63\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.808804 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-inventory\") pod \"d0c74792-a006-4752-8403-70947f6fec63\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.808910 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-ssh-key-openstack-edpm-ipam\") pod \"d0c74792-a006-4752-8403-70947f6fec63\" (UID: \"d0c74792-a006-4752-8403-70947f6fec63\") " Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.815040 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0c74792-a006-4752-8403-70947f6fec63-kube-api-access-grgn8" (OuterVolumeSpecName: "kube-api-access-grgn8") pod "d0c74792-a006-4752-8403-70947f6fec63" (UID: "d0c74792-a006-4752-8403-70947f6fec63"). InnerVolumeSpecName "kube-api-access-grgn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.831314 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d0c74792-a006-4752-8403-70947f6fec63" (UID: "d0c74792-a006-4752-8403-70947f6fec63"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.837011 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-inventory" (OuterVolumeSpecName: "inventory") pod "d0c74792-a006-4752-8403-70947f6fec63" (UID: "d0c74792-a006-4752-8403-70947f6fec63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.913548 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grgn8\" (UniqueName: \"kubernetes.io/projected/d0c74792-a006-4752-8403-70947f6fec63-kube-api-access-grgn8\") on node \"crc\" DevicePath \"\"" Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.913584 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:07:07 crc kubenswrapper[4492]: I0220 07:07:07.913601 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d0c74792-a006-4752-8403-70947f6fec63-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.241934 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" event={"ID":"d0c74792-a006-4752-8403-70947f6fec63","Type":"ContainerDied","Data":"e5b1abd96c168e366c2eb5a9f50873a21531055dd6d81abb5b77cd12b17ce048"} Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.241995 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5b1abd96c168e366c2eb5a9f50873a21531055dd6d81abb5b77cd12b17ce048" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.242011 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326371 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj"] Feb 20 07:07:08 crc kubenswrapper[4492]: E0220 07:07:08.326808 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="extract-utilities" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326828 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="extract-utilities" Feb 20 07:07:08 crc kubenswrapper[4492]: E0220 07:07:08.326850 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="registry-server" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326857 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="registry-server" Feb 20 07:07:08 crc kubenswrapper[4492]: E0220 07:07:08.326869 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="extract-utilities" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326875 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="extract-utilities" Feb 20 07:07:08 crc kubenswrapper[4492]: E0220 07:07:08.326893 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0c74792-a006-4752-8403-70947f6fec63" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326901 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0c74792-a006-4752-8403-70947f6fec63" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 07:07:08 crc kubenswrapper[4492]: E0220 07:07:08.326918 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="registry-server" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326923 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="registry-server" Feb 20 07:07:08 crc kubenswrapper[4492]: E0220 07:07:08.326936 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="extract-content" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326941 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="extract-content" Feb 20 07:07:08 crc kubenswrapper[4492]: E0220 07:07:08.326953 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="extract-content" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.326960 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="extract-content" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.327137 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="73fc93b1-fb18-4271-af6f-ad1278738fdc" containerName="registry-server" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.327154 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="65910b40-8daf-4cbd-91ca-e5453ec08bba" containerName="registry-server" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.327169 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0c74792-a006-4752-8403-70947f6fec63" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.327854 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.332975 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.333076 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.333394 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.337194 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.338799 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj"] Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.436992 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.437384 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wwbr\" (UniqueName: \"kubernetes.io/projected/78fa6f40-87e3-4c44-b233-0a83c8e8d429-kube-api-access-2wwbr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.437503 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.538749 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.539086 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wwbr\" (UniqueName: \"kubernetes.io/projected/78fa6f40-87e3-4c44-b233-0a83c8e8d429-kube-api-access-2wwbr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.539226 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.543543 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.544191 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.558584 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wwbr\" (UniqueName: \"kubernetes.io/projected/78fa6f40-87e3-4c44-b233-0a83c8e8d429-kube-api-access-2wwbr\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:08 crc kubenswrapper[4492]: I0220 07:07:08.642082 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:07:09 crc kubenswrapper[4492]: I0220 07:07:09.139127 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj"] Feb 20 07:07:09 crc kubenswrapper[4492]: I0220 07:07:09.254304 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" event={"ID":"78fa6f40-87e3-4c44-b233-0a83c8e8d429","Type":"ContainerStarted","Data":"c9d8c4ffd2b09e55140fa9cd697ad1cfbe9c0848c34ab64b7141ac4cda397056"} Feb 20 07:07:10 crc kubenswrapper[4492]: I0220 07:07:10.269169 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" event={"ID":"78fa6f40-87e3-4c44-b233-0a83c8e8d429","Type":"ContainerStarted","Data":"57aafceb1f897e299771737a48d43bc75304742d7ea321aaa177c3d2c8e3003a"} Feb 20 07:07:10 crc kubenswrapper[4492]: I0220 07:07:10.300059 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" podStartSLOduration=1.807942677 podStartE2EDuration="2.300035303s" podCreationTimestamp="2026-02-20 07:07:08 +0000 UTC" firstStartedPulling="2026-02-20 07:07:09.141123117 +0000 UTC m=+1585.912412096" lastFinishedPulling="2026-02-20 07:07:09.633215744 +0000 UTC m=+1586.404504722" observedRunningTime="2026-02-20 07:07:10.284437469 +0000 UTC m=+1587.055726447" watchObservedRunningTime="2026-02-20 07:07:10.300035303 +0000 UTC m=+1587.071324282" Feb 20 07:07:11 crc kubenswrapper[4492]: I0220 07:07:11.558694 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:07:11 crc kubenswrapper[4492]: E0220 07:07:11.559209 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:07:18 crc kubenswrapper[4492]: I0220 07:07:18.043601 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-d79gm"] Feb 20 07:07:18 crc kubenswrapper[4492]: I0220 07:07:18.053318 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-d79gm"] Feb 20 07:07:19 crc kubenswrapper[4492]: I0220 07:07:19.030128 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-qqhvc"] Feb 20 07:07:19 crc kubenswrapper[4492]: I0220 07:07:19.038516 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-qqhvc"] Feb 20 07:07:19 crc kubenswrapper[4492]: I0220 07:07:19.569604 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05565eb0-90db-4ec7-b646-db528fca40ad" path="/var/lib/kubelet/pods/05565eb0-90db-4ec7-b646-db528fca40ad/volumes" Feb 20 07:07:19 crc kubenswrapper[4492]: I0220 07:07:19.570955 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ed151d-241b-465d-8b5d-d97fe51015df" path="/var/lib/kubelet/pods/82ed151d-241b-465d-8b5d-d97fe51015df/volumes" Feb 20 07:07:26 crc kubenswrapper[4492]: I0220 07:07:26.557601 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:07:26 crc kubenswrapper[4492]: E0220 07:07:26.558261 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.393604 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bt8d5"] Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.396328 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.406552 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bt8d5"] Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.532090 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-utilities\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.532187 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-catalog-content\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.532346 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t552\" (UniqueName: \"kubernetes.io/projected/0feb69bc-506c-4f53-a7ca-0a443524ae64-kube-api-access-5t552\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.636593 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-utilities\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.636725 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-catalog-content\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.637059 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t552\" (UniqueName: \"kubernetes.io/projected/0feb69bc-506c-4f53-a7ca-0a443524ae64-kube-api-access-5t552\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.637111 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-utilities\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.637882 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-catalog-content\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.664928 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t552\" (UniqueName: \"kubernetes.io/projected/0feb69bc-506c-4f53-a7ca-0a443524ae64-kube-api-access-5t552\") pod \"redhat-operators-bt8d5\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:34 crc kubenswrapper[4492]: I0220 07:07:34.730284 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:35 crc kubenswrapper[4492]: I0220 07:07:35.008033 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bt8d5"] Feb 20 07:07:35 crc kubenswrapper[4492]: I0220 07:07:35.551686 4492 generic.go:334] "Generic (PLEG): container finished" podID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerID="8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262" exitCode=0 Feb 20 07:07:35 crc kubenswrapper[4492]: I0220 07:07:35.551997 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bt8d5" event={"ID":"0feb69bc-506c-4f53-a7ca-0a443524ae64","Type":"ContainerDied","Data":"8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262"} Feb 20 07:07:35 crc kubenswrapper[4492]: I0220 07:07:35.552030 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bt8d5" event={"ID":"0feb69bc-506c-4f53-a7ca-0a443524ae64","Type":"ContainerStarted","Data":"a20f3c237863b76934ddd235dd21ea46de70df6d94a83199956c410efb57c23c"} Feb 20 07:07:36 crc kubenswrapper[4492]: I0220 07:07:36.564706 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bt8d5" event={"ID":"0feb69bc-506c-4f53-a7ca-0a443524ae64","Type":"ContainerStarted","Data":"b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948"} Feb 20 07:07:39 crc kubenswrapper[4492]: I0220 07:07:39.606822 4492 generic.go:334] "Generic (PLEG): container finished" podID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerID="b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948" exitCode=0 Feb 20 07:07:39 crc kubenswrapper[4492]: I0220 07:07:39.607313 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bt8d5" event={"ID":"0feb69bc-506c-4f53-a7ca-0a443524ae64","Type":"ContainerDied","Data":"b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948"} Feb 20 07:07:40 crc kubenswrapper[4492]: I0220 07:07:40.619815 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bt8d5" event={"ID":"0feb69bc-506c-4f53-a7ca-0a443524ae64","Type":"ContainerStarted","Data":"fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8"} Feb 20 07:07:41 crc kubenswrapper[4492]: I0220 07:07:41.557841 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:07:41 crc kubenswrapper[4492]: E0220 07:07:41.558080 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:07:44 crc kubenswrapper[4492]: I0220 07:07:44.731159 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:44 crc kubenswrapper[4492]: I0220 07:07:44.731994 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:07:45 crc kubenswrapper[4492]: I0220 07:07:45.780683 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bt8d5" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="registry-server" probeResult="failure" output=< Feb 20 07:07:45 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:07:45 crc kubenswrapper[4492]: > Feb 20 07:07:54 crc kubenswrapper[4492]: I0220 07:07:54.557508 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:07:54 crc kubenswrapper[4492]: E0220 07:07:54.558643 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:07:55 crc kubenswrapper[4492]: I0220 07:07:55.773154 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bt8d5" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="registry-server" probeResult="failure" output=< Feb 20 07:07:55 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:07:55 crc kubenswrapper[4492]: > Feb 20 07:08:04 crc kubenswrapper[4492]: I0220 07:08:04.774560 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:08:04 crc kubenswrapper[4492]: I0220 07:08:04.802502 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bt8d5" podStartSLOduration=26.283307748 podStartE2EDuration="30.802461209s" podCreationTimestamp="2026-02-20 07:07:34 +0000 UTC" firstStartedPulling="2026-02-20 07:07:35.55337963 +0000 UTC m=+1612.324668608" lastFinishedPulling="2026-02-20 07:07:40.072533091 +0000 UTC m=+1616.843822069" observedRunningTime="2026-02-20 07:07:40.642719911 +0000 UTC m=+1617.414008889" watchObservedRunningTime="2026-02-20 07:08:04.802461209 +0000 UTC m=+1641.573750187" Feb 20 07:08:04 crc kubenswrapper[4492]: I0220 07:08:04.818428 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:08:05 crc kubenswrapper[4492]: I0220 07:08:05.601677 4492 scope.go:117] "RemoveContainer" containerID="5243939417ea46c026a621e02fc03fc327f97535c2337c2c6f66159db20dd056" Feb 20 07:08:05 crc kubenswrapper[4492]: I0220 07:08:05.610697 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bt8d5"] Feb 20 07:08:05 crc kubenswrapper[4492]: I0220 07:08:05.651461 4492 scope.go:117] "RemoveContainer" containerID="266d6cf492d9d445c03d3eeb2f23bcd26684b3384081232720c9f7aa0fd70dc6" Feb 20 07:08:05 crc kubenswrapper[4492]: I0220 07:08:05.705862 4492 scope.go:117] "RemoveContainer" containerID="9dff1702bfc3db58505276fe1faa25079b995045b18caee73490500f3e3d21b5" Feb 20 07:08:05 crc kubenswrapper[4492]: I0220 07:08:05.749045 4492 scope.go:117] "RemoveContainer" containerID="713a78e6ee4d4e6d901fc9c60d99ba58b4619f1476820ee39be63f437b86d1d1" Feb 20 07:08:05 crc kubenswrapper[4492]: I0220 07:08:05.810205 4492 scope.go:117] "RemoveContainer" containerID="10a248028db76ad41a2932cdc1b43904df882eec0fcdf41f9a8c10691b38f7b5" Feb 20 07:08:05 crc kubenswrapper[4492]: I0220 07:08:05.950425 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bt8d5" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="registry-server" containerID="cri-o://fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8" gracePeriod=2 Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.347205 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.512962 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-catalog-content\") pod \"0feb69bc-506c-4f53-a7ca-0a443524ae64\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.513176 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t552\" (UniqueName: \"kubernetes.io/projected/0feb69bc-506c-4f53-a7ca-0a443524ae64-kube-api-access-5t552\") pod \"0feb69bc-506c-4f53-a7ca-0a443524ae64\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.513257 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-utilities\") pod \"0feb69bc-506c-4f53-a7ca-0a443524ae64\" (UID: \"0feb69bc-506c-4f53-a7ca-0a443524ae64\") " Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.514359 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-utilities" (OuterVolumeSpecName: "utilities") pod "0feb69bc-506c-4f53-a7ca-0a443524ae64" (UID: "0feb69bc-506c-4f53-a7ca-0a443524ae64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.521342 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0feb69bc-506c-4f53-a7ca-0a443524ae64-kube-api-access-5t552" (OuterVolumeSpecName: "kube-api-access-5t552") pod "0feb69bc-506c-4f53-a7ca-0a443524ae64" (UID: "0feb69bc-506c-4f53-a7ca-0a443524ae64"). InnerVolumeSpecName "kube-api-access-5t552". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.602219 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0feb69bc-506c-4f53-a7ca-0a443524ae64" (UID: "0feb69bc-506c-4f53-a7ca-0a443524ae64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.616669 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.616719 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t552\" (UniqueName: \"kubernetes.io/projected/0feb69bc-506c-4f53-a7ca-0a443524ae64-kube-api-access-5t552\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.616746 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0feb69bc-506c-4f53-a7ca-0a443524ae64-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.960433 4492 generic.go:334] "Generic (PLEG): container finished" podID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerID="fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8" exitCode=0 Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.960538 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bt8d5" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.960538 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bt8d5" event={"ID":"0feb69bc-506c-4f53-a7ca-0a443524ae64","Type":"ContainerDied","Data":"fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8"} Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.960674 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bt8d5" event={"ID":"0feb69bc-506c-4f53-a7ca-0a443524ae64","Type":"ContainerDied","Data":"a20f3c237863b76934ddd235dd21ea46de70df6d94a83199956c410efb57c23c"} Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.960704 4492 scope.go:117] "RemoveContainer" containerID="fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.984747 4492 scope.go:117] "RemoveContainer" containerID="b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948" Feb 20 07:08:06 crc kubenswrapper[4492]: I0220 07:08:06.994190 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bt8d5"] Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.007926 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bt8d5"] Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.015385 4492 scope.go:117] "RemoveContainer" containerID="8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262" Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.060751 4492 scope.go:117] "RemoveContainer" containerID="fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8" Feb 20 07:08:07 crc kubenswrapper[4492]: E0220 07:08:07.061678 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8\": container with ID starting with fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8 not found: ID does not exist" containerID="fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8" Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.061936 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8"} err="failed to get container status \"fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8\": rpc error: code = NotFound desc = could not find container \"fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8\": container with ID starting with fbcda34f485a4b6347feb26270b19569d924ec343056c70dd006064dd709d2b8 not found: ID does not exist" Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.062541 4492 scope.go:117] "RemoveContainer" containerID="b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948" Feb 20 07:08:07 crc kubenswrapper[4492]: E0220 07:08:07.062823 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948\": container with ID starting with b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948 not found: ID does not exist" containerID="b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948" Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.062867 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948"} err="failed to get container status \"b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948\": rpc error: code = NotFound desc = could not find container \"b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948\": container with ID starting with b17ab0017d98abec651a4b93371db5af6972d4ae2659344cd201fb5574293948 not found: ID does not exist" Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.062897 4492 scope.go:117] "RemoveContainer" containerID="8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262" Feb 20 07:08:07 crc kubenswrapper[4492]: E0220 07:08:07.063357 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262\": container with ID starting with 8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262 not found: ID does not exist" containerID="8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262" Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.063380 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262"} err="failed to get container status \"8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262\": rpc error: code = NotFound desc = could not find container \"8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262\": container with ID starting with 8b790c2f7cb0089c1fae90f8e7eb69c2c18c9adc087afe45e2739c1192792262 not found: ID does not exist" Feb 20 07:08:07 crc kubenswrapper[4492]: I0220 07:08:07.564738 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" path="/var/lib/kubelet/pods/0feb69bc-506c-4f53-a7ca-0a443524ae64/volumes" Feb 20 07:08:09 crc kubenswrapper[4492]: I0220 07:08:09.558544 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:08:09 crc kubenswrapper[4492]: E0220 07:08:09.559147 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:08:11 crc kubenswrapper[4492]: I0220 07:08:11.032453 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-g5r5h"] Feb 20 07:08:11 crc kubenswrapper[4492]: I0220 07:08:11.038998 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-g5r5h"] Feb 20 07:08:11 crc kubenswrapper[4492]: I0220 07:08:11.565056 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b92062-3c5c-40e0-9ccf-72e9f896a785" path="/var/lib/kubelet/pods/f6b92062-3c5c-40e0-9ccf-72e9f896a785/volumes" Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.042153 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jwgdt"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.048328 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f823-account-create-update-hkhrt"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.054146 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-be0c-account-create-update-ss6pn"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.059867 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ea34-account-create-update-f96zd"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.065615 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jwgdt"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.075588 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-6tdxw"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.081267 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-be0c-account-create-update-ss6pn"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.091343 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f823-account-create-update-hkhrt"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.095740 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ea34-account-create-update-f96zd"] Feb 20 07:08:12 crc kubenswrapper[4492]: I0220 07:08:12.101967 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-6tdxw"] Feb 20 07:08:13 crc kubenswrapper[4492]: I0220 07:08:13.566184 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081589a6-39f6-4bb6-970d-194e8fa6bc42" path="/var/lib/kubelet/pods/081589a6-39f6-4bb6-970d-194e8fa6bc42/volumes" Feb 20 07:08:13 crc kubenswrapper[4492]: I0220 07:08:13.568049 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33867d4d-53f0-4f25-b61b-f02a98c3b047" path="/var/lib/kubelet/pods/33867d4d-53f0-4f25-b61b-f02a98c3b047/volumes" Feb 20 07:08:13 crc kubenswrapper[4492]: I0220 07:08:13.568891 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c405394d-8e8e-4c21-ae6e-9de9c630c4c3" path="/var/lib/kubelet/pods/c405394d-8e8e-4c21-ae6e-9de9c630c4c3/volumes" Feb 20 07:08:13 crc kubenswrapper[4492]: I0220 07:08:13.569690 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e96057e9-436d-4a08-bf9d-6ae2de326502" path="/var/lib/kubelet/pods/e96057e9-436d-4a08-bf9d-6ae2de326502/volumes" Feb 20 07:08:13 crc kubenswrapper[4492]: I0220 07:08:13.571122 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88957ee-9211-4a0e-98f0-0a609cbe03bb" path="/var/lib/kubelet/pods/f88957ee-9211-4a0e-98f0-0a609cbe03bb/volumes" Feb 20 07:08:14 crc kubenswrapper[4492]: I0220 07:08:14.020706 4492 generic.go:334] "Generic (PLEG): container finished" podID="78fa6f40-87e3-4c44-b233-0a83c8e8d429" containerID="57aafceb1f897e299771737a48d43bc75304742d7ea321aaa177c3d2c8e3003a" exitCode=0 Feb 20 07:08:14 crc kubenswrapper[4492]: I0220 07:08:14.020787 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" event={"ID":"78fa6f40-87e3-4c44-b233-0a83c8e8d429","Type":"ContainerDied","Data":"57aafceb1f897e299771737a48d43bc75304742d7ea321aaa177c3d2c8e3003a"} Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.362798 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.494889 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-ssh-key-openstack-edpm-ipam\") pod \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.494962 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-inventory\") pod \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.495028 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wwbr\" (UniqueName: \"kubernetes.io/projected/78fa6f40-87e3-4c44-b233-0a83c8e8d429-kube-api-access-2wwbr\") pod \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\" (UID: \"78fa6f40-87e3-4c44-b233-0a83c8e8d429\") " Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.506736 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78fa6f40-87e3-4c44-b233-0a83c8e8d429-kube-api-access-2wwbr" (OuterVolumeSpecName: "kube-api-access-2wwbr") pod "78fa6f40-87e3-4c44-b233-0a83c8e8d429" (UID: "78fa6f40-87e3-4c44-b233-0a83c8e8d429"). InnerVolumeSpecName "kube-api-access-2wwbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.517234 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-inventory" (OuterVolumeSpecName: "inventory") pod "78fa6f40-87e3-4c44-b233-0a83c8e8d429" (UID: "78fa6f40-87e3-4c44-b233-0a83c8e8d429"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.522129 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "78fa6f40-87e3-4c44-b233-0a83c8e8d429" (UID: "78fa6f40-87e3-4c44-b233-0a83c8e8d429"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.597081 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.597346 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78fa6f40-87e3-4c44-b233-0a83c8e8d429-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:15 crc kubenswrapper[4492]: I0220 07:08:15.597358 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wwbr\" (UniqueName: \"kubernetes.io/projected/78fa6f40-87e3-4c44-b233-0a83c8e8d429-kube-api-access-2wwbr\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.035733 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" event={"ID":"78fa6f40-87e3-4c44-b233-0a83c8e8d429","Type":"ContainerDied","Data":"c9d8c4ffd2b09e55140fa9cd697ad1cfbe9c0848c34ab64b7141ac4cda397056"} Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.035783 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9d8c4ffd2b09e55140fa9cd697ad1cfbe9c0848c34ab64b7141ac4cda397056" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.035846 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.111049 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg"] Feb 20 07:08:16 crc kubenswrapper[4492]: E0220 07:08:16.112026 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="extract-content" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.112049 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="extract-content" Feb 20 07:08:16 crc kubenswrapper[4492]: E0220 07:08:16.112079 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="registry-server" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.112086 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="registry-server" Feb 20 07:08:16 crc kubenswrapper[4492]: E0220 07:08:16.112110 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="extract-utilities" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.112121 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="extract-utilities" Feb 20 07:08:16 crc kubenswrapper[4492]: E0220 07:08:16.112137 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78fa6f40-87e3-4c44-b233-0a83c8e8d429" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.112144 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="78fa6f40-87e3-4c44-b233-0a83c8e8d429" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.112549 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0feb69bc-506c-4f53-a7ca-0a443524ae64" containerName="registry-server" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.112573 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="78fa6f40-87e3-4c44-b233-0a83c8e8d429" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.113825 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.119562 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.120035 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.120318 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.120613 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.163208 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg"] Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.313749 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.313831 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnjgr\" (UniqueName: \"kubernetes.io/projected/62a602f6-1225-4dc5-9959-8521e0d07631-kube-api-access-qnjgr\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.313901 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.415656 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.415728 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnjgr\" (UniqueName: \"kubernetes.io/projected/62a602f6-1225-4dc5-9959-8521e0d07631-kube-api-access-qnjgr\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.415784 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.420901 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.421639 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.432061 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnjgr\" (UniqueName: \"kubernetes.io/projected/62a602f6-1225-4dc5-9959-8521e0d07631-kube-api-access-qnjgr\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.442944 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:16 crc kubenswrapper[4492]: I0220 07:08:16.909647 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg"] Feb 20 07:08:17 crc kubenswrapper[4492]: I0220 07:08:17.045259 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" event={"ID":"62a602f6-1225-4dc5-9959-8521e0d07631","Type":"ContainerStarted","Data":"227113f6f531c21f6b5ea75ffaa39b6de15f6cf5b8d80f86ddb183fa56dce33b"} Feb 20 07:08:18 crc kubenswrapper[4492]: I0220 07:08:18.053566 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" event={"ID":"62a602f6-1225-4dc5-9959-8521e0d07631","Type":"ContainerStarted","Data":"25d58cebcde2d650482daa53a5399e62027a5e9993c76f0e7831a72e5c6dba37"} Feb 20 07:08:18 crc kubenswrapper[4492]: I0220 07:08:18.071009 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" podStartSLOduration=1.5840504869999998 podStartE2EDuration="2.070994057s" podCreationTimestamp="2026-02-20 07:08:16 +0000 UTC" firstStartedPulling="2026-02-20 07:08:16.912525698 +0000 UTC m=+1653.683814676" lastFinishedPulling="2026-02-20 07:08:17.399469268 +0000 UTC m=+1654.170758246" observedRunningTime="2026-02-20 07:08:18.06371744 +0000 UTC m=+1654.835006418" watchObservedRunningTime="2026-02-20 07:08:18.070994057 +0000 UTC m=+1654.842283035" Feb 20 07:08:21 crc kubenswrapper[4492]: I0220 07:08:21.556969 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:08:21 crc kubenswrapper[4492]: E0220 07:08:21.557783 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:08:22 crc kubenswrapper[4492]: I0220 07:08:22.085358 4492 generic.go:334] "Generic (PLEG): container finished" podID="62a602f6-1225-4dc5-9959-8521e0d07631" containerID="25d58cebcde2d650482daa53a5399e62027a5e9993c76f0e7831a72e5c6dba37" exitCode=0 Feb 20 07:08:22 crc kubenswrapper[4492]: I0220 07:08:22.085401 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" event={"ID":"62a602f6-1225-4dc5-9959-8521e0d07631","Type":"ContainerDied","Data":"25d58cebcde2d650482daa53a5399e62027a5e9993c76f0e7831a72e5c6dba37"} Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.402567 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.534587 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-ssh-key-openstack-edpm-ipam\") pod \"62a602f6-1225-4dc5-9959-8521e0d07631\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.534705 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-inventory\") pod \"62a602f6-1225-4dc5-9959-8521e0d07631\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.535240 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnjgr\" (UniqueName: \"kubernetes.io/projected/62a602f6-1225-4dc5-9959-8521e0d07631-kube-api-access-qnjgr\") pod \"62a602f6-1225-4dc5-9959-8521e0d07631\" (UID: \"62a602f6-1225-4dc5-9959-8521e0d07631\") " Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.545018 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a602f6-1225-4dc5-9959-8521e0d07631-kube-api-access-qnjgr" (OuterVolumeSpecName: "kube-api-access-qnjgr") pod "62a602f6-1225-4dc5-9959-8521e0d07631" (UID: "62a602f6-1225-4dc5-9959-8521e0d07631"). InnerVolumeSpecName "kube-api-access-qnjgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.556622 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-inventory" (OuterVolumeSpecName: "inventory") pod "62a602f6-1225-4dc5-9959-8521e0d07631" (UID: "62a602f6-1225-4dc5-9959-8521e0d07631"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.563534 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "62a602f6-1225-4dc5-9959-8521e0d07631" (UID: "62a602f6-1225-4dc5-9959-8521e0d07631"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.638328 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.638360 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62a602f6-1225-4dc5-9959-8521e0d07631-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:23 crc kubenswrapper[4492]: I0220 07:08:23.638373 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnjgr\" (UniqueName: \"kubernetes.io/projected/62a602f6-1225-4dc5-9959-8521e0d07631-kube-api-access-qnjgr\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.102552 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" event={"ID":"62a602f6-1225-4dc5-9959-8521e0d07631","Type":"ContainerDied","Data":"227113f6f531c21f6b5ea75ffaa39b6de15f6cf5b8d80f86ddb183fa56dce33b"} Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.102597 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="227113f6f531c21f6b5ea75ffaa39b6de15f6cf5b8d80f86ddb183fa56dce33b" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.102620 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.159365 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62"] Feb 20 07:08:24 crc kubenswrapper[4492]: E0220 07:08:24.159746 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a602f6-1225-4dc5-9959-8521e0d07631" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.159764 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a602f6-1225-4dc5-9959-8521e0d07631" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.159934 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="62a602f6-1225-4dc5-9959-8521e0d07631" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.160618 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.162693 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.162952 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.163074 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.163975 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.168074 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62"] Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.248979 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.249058 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lz6h\" (UniqueName: \"kubernetes.io/projected/cccd9a33-167d-41ec-b6ab-833d1273f12a-kube-api-access-4lz6h\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.249199 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.351509 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.351887 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lz6h\" (UniqueName: \"kubernetes.io/projected/cccd9a33-167d-41ec-b6ab-833d1273f12a-kube-api-access-4lz6h\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.352007 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.357958 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.357958 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.369258 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lz6h\" (UniqueName: \"kubernetes.io/projected/cccd9a33-167d-41ec-b6ab-833d1273f12a-kube-api-access-4lz6h\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mdx62\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.474140 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:24 crc kubenswrapper[4492]: I0220 07:08:24.936170 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62"] Feb 20 07:08:25 crc kubenswrapper[4492]: I0220 07:08:25.111526 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" event={"ID":"cccd9a33-167d-41ec-b6ab-833d1273f12a","Type":"ContainerStarted","Data":"7e9622401e77a6165e61b7e7e6bb65ec40642178735089ba2c424bec3d21fde7"} Feb 20 07:08:26 crc kubenswrapper[4492]: I0220 07:08:26.122167 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" event={"ID":"cccd9a33-167d-41ec-b6ab-833d1273f12a","Type":"ContainerStarted","Data":"a75196fd3f5c050e5b4a28423dee350bb1cf4a58b155c938b77d998a3fa7c03c"} Feb 20 07:08:26 crc kubenswrapper[4492]: I0220 07:08:26.142972 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" podStartSLOduration=1.6417789489999999 podStartE2EDuration="2.142958747s" podCreationTimestamp="2026-02-20 07:08:24 +0000 UTC" firstStartedPulling="2026-02-20 07:08:24.94333494 +0000 UTC m=+1661.714623919" lastFinishedPulling="2026-02-20 07:08:25.444514739 +0000 UTC m=+1662.215803717" observedRunningTime="2026-02-20 07:08:26.136711139 +0000 UTC m=+1662.908000117" watchObservedRunningTime="2026-02-20 07:08:26.142958747 +0000 UTC m=+1662.914247725" Feb 20 07:08:35 crc kubenswrapper[4492]: I0220 07:08:35.556571 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:08:35 crc kubenswrapper[4492]: E0220 07:08:35.557317 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:08:45 crc kubenswrapper[4492]: I0220 07:08:45.032108 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bsh7r"] Feb 20 07:08:45 crc kubenswrapper[4492]: I0220 07:08:45.050860 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-bsh7r"] Feb 20 07:08:45 crc kubenswrapper[4492]: I0220 07:08:45.566401 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d97b093e-5584-4662-a976-215ada9ef003" path="/var/lib/kubelet/pods/d97b093e-5584-4662-a976-215ada9ef003/volumes" Feb 20 07:08:48 crc kubenswrapper[4492]: I0220 07:08:48.556413 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:08:48 crc kubenswrapper[4492]: E0220 07:08:48.556952 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:08:53 crc kubenswrapper[4492]: I0220 07:08:53.357491 4492 generic.go:334] "Generic (PLEG): container finished" podID="cccd9a33-167d-41ec-b6ab-833d1273f12a" containerID="a75196fd3f5c050e5b4a28423dee350bb1cf4a58b155c938b77d998a3fa7c03c" exitCode=0 Feb 20 07:08:53 crc kubenswrapper[4492]: I0220 07:08:53.357574 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" event={"ID":"cccd9a33-167d-41ec-b6ab-833d1273f12a","Type":"ContainerDied","Data":"a75196fd3f5c050e5b4a28423dee350bb1cf4a58b155c938b77d998a3fa7c03c"} Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.729219 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.843683 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-inventory\") pod \"cccd9a33-167d-41ec-b6ab-833d1273f12a\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.843875 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-ssh-key-openstack-edpm-ipam\") pod \"cccd9a33-167d-41ec-b6ab-833d1273f12a\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.843969 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lz6h\" (UniqueName: \"kubernetes.io/projected/cccd9a33-167d-41ec-b6ab-833d1273f12a-kube-api-access-4lz6h\") pod \"cccd9a33-167d-41ec-b6ab-833d1273f12a\" (UID: \"cccd9a33-167d-41ec-b6ab-833d1273f12a\") " Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.852409 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cccd9a33-167d-41ec-b6ab-833d1273f12a-kube-api-access-4lz6h" (OuterVolumeSpecName: "kube-api-access-4lz6h") pod "cccd9a33-167d-41ec-b6ab-833d1273f12a" (UID: "cccd9a33-167d-41ec-b6ab-833d1273f12a"). InnerVolumeSpecName "kube-api-access-4lz6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.872866 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cccd9a33-167d-41ec-b6ab-833d1273f12a" (UID: "cccd9a33-167d-41ec-b6ab-833d1273f12a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.873416 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-inventory" (OuterVolumeSpecName: "inventory") pod "cccd9a33-167d-41ec-b6ab-833d1273f12a" (UID: "cccd9a33-167d-41ec-b6ab-833d1273f12a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.947102 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.947194 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cccd9a33-167d-41ec-b6ab-833d1273f12a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:54 crc kubenswrapper[4492]: I0220 07:08:54.947254 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lz6h\" (UniqueName: \"kubernetes.io/projected/cccd9a33-167d-41ec-b6ab-833d1273f12a-kube-api-access-4lz6h\") on node \"crc\" DevicePath \"\"" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.381830 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" event={"ID":"cccd9a33-167d-41ec-b6ab-833d1273f12a","Type":"ContainerDied","Data":"7e9622401e77a6165e61b7e7e6bb65ec40642178735089ba2c424bec3d21fde7"} Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.381878 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e9622401e77a6165e61b7e7e6bb65ec40642178735089ba2c424bec3d21fde7" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.381937 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mdx62" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.465286 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd"] Feb 20 07:08:55 crc kubenswrapper[4492]: E0220 07:08:55.466085 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cccd9a33-167d-41ec-b6ab-833d1273f12a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.466104 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="cccd9a33-167d-41ec-b6ab-833d1273f12a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.466307 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="cccd9a33-167d-41ec-b6ab-833d1273f12a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.467121 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.470066 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.470131 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.470239 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.470556 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.484909 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd"] Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.659270 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.660637 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.660694 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br2lj\" (UniqueName: \"kubernetes.io/projected/719a407b-3228-4bbc-a600-c08f555ba556-kube-api-access-br2lj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.762100 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.762163 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br2lj\" (UniqueName: \"kubernetes.io/projected/719a407b-3228-4bbc-a600-c08f555ba556-kube-api-access-br2lj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.762234 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.767289 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.768878 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.776304 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br2lj\" (UniqueName: \"kubernetes.io/projected/719a407b-3228-4bbc-a600-c08f555ba556-kube-api-access-br2lj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:55 crc kubenswrapper[4492]: I0220 07:08:55.786397 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:08:56 crc kubenswrapper[4492]: I0220 07:08:56.249136 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd"] Feb 20 07:08:56 crc kubenswrapper[4492]: I0220 07:08:56.389737 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" event={"ID":"719a407b-3228-4bbc-a600-c08f555ba556","Type":"ContainerStarted","Data":"be2c9ee9a0593650f89de4c5d04e373cf4f07b9525a457ce208dd1a71ca25a68"} Feb 20 07:08:57 crc kubenswrapper[4492]: I0220 07:08:57.397799 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" event={"ID":"719a407b-3228-4bbc-a600-c08f555ba556","Type":"ContainerStarted","Data":"3dd47efb94381e0eede39e19c59bdb6c288a214c872ccaf5992169e564fb4b5e"} Feb 20 07:08:57 crc kubenswrapper[4492]: I0220 07:08:57.412041 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" podStartSLOduration=1.9431821440000001 podStartE2EDuration="2.412030548s" podCreationTimestamp="2026-02-20 07:08:55 +0000 UTC" firstStartedPulling="2026-02-20 07:08:56.244927423 +0000 UTC m=+1693.016216401" lastFinishedPulling="2026-02-20 07:08:56.713775826 +0000 UTC m=+1693.485064805" observedRunningTime="2026-02-20 07:08:57.410364968 +0000 UTC m=+1694.181653946" watchObservedRunningTime="2026-02-20 07:08:57.412030548 +0000 UTC m=+1694.183319526" Feb 20 07:09:03 crc kubenswrapper[4492]: I0220 07:09:03.561650 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:09:03 crc kubenswrapper[4492]: E0220 07:09:03.562374 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:09:06 crc kubenswrapper[4492]: I0220 07:09:06.035835 4492 scope.go:117] "RemoveContainer" containerID="7c135289b2f6d8cf45477101b9b80c6d3e0e29428418687222ef99ee34b7de3f" Feb 20 07:09:06 crc kubenswrapper[4492]: I0220 07:09:06.057467 4492 scope.go:117] "RemoveContainer" containerID="c217337af4e2aca3a2b02e87f11936f17070bb9e635a3363dfcce2813ce50543" Feb 20 07:09:06 crc kubenswrapper[4492]: I0220 07:09:06.117956 4492 scope.go:117] "RemoveContainer" containerID="42d5e00fa14a4d8a12fcabf40684c00c84bd593c160f043fbcd340afa3f4128c" Feb 20 07:09:06 crc kubenswrapper[4492]: I0220 07:09:06.147583 4492 scope.go:117] "RemoveContainer" containerID="897fe67f3c9ef560a51c7c60257af140e5abb4878b990c0331400b8211873e6d" Feb 20 07:09:06 crc kubenswrapper[4492]: I0220 07:09:06.180816 4492 scope.go:117] "RemoveContainer" containerID="7f0f8e79841d1114a5bd817e65c94bd327be8a4c19392bc146b97a3c66ec3466" Feb 20 07:09:06 crc kubenswrapper[4492]: I0220 07:09:06.216499 4492 scope.go:117] "RemoveContainer" containerID="bcfbae3e901b7ae2b74b583bb82a015b944ad3450116bc17d5528d80d396cea3" Feb 20 07:09:06 crc kubenswrapper[4492]: I0220 07:09:06.250752 4492 scope.go:117] "RemoveContainer" containerID="e2b485e3a95d0f7b13fdfe393188e41df1e0ddaadbf4bfa4485f53d2d8f7543f" Feb 20 07:09:08 crc kubenswrapper[4492]: I0220 07:09:08.042276 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rj7nh"] Feb 20 07:09:08 crc kubenswrapper[4492]: I0220 07:09:08.057564 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-rj7nh"] Feb 20 07:09:09 crc kubenswrapper[4492]: I0220 07:09:09.029566 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-bt5pz"] Feb 20 07:09:09 crc kubenswrapper[4492]: I0220 07:09:09.039945 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-bt5pz"] Feb 20 07:09:09 crc kubenswrapper[4492]: I0220 07:09:09.567755 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4c9cd39-beac-4a57-ba33-ae3f41187e8e" path="/var/lib/kubelet/pods/a4c9cd39-beac-4a57-ba33-ae3f41187e8e/volumes" Feb 20 07:09:09 crc kubenswrapper[4492]: I0220 07:09:09.568757 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0a2bbcf-c2d7-47b7-873d-4b93d85c963a" path="/var/lib/kubelet/pods/f0a2bbcf-c2d7-47b7-873d-4b93d85c963a/volumes" Feb 20 07:09:17 crc kubenswrapper[4492]: I0220 07:09:17.557202 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:09:17 crc kubenswrapper[4492]: E0220 07:09:17.558136 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:09:32 crc kubenswrapper[4492]: I0220 07:09:32.556798 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:09:32 crc kubenswrapper[4492]: E0220 07:09:32.557710 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:09:33 crc kubenswrapper[4492]: E0220 07:09:33.081892 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod719a407b_3228_4bbc_a600_c08f555ba556.slice/crio-conmon-3dd47efb94381e0eede39e19c59bdb6c288a214c872ccaf5992169e564fb4b5e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod719a407b_3228_4bbc_a600_c08f555ba556.slice/crio-3dd47efb94381e0eede39e19c59bdb6c288a214c872ccaf5992169e564fb4b5e.scope\": RecentStats: unable to find data in memory cache]" Feb 20 07:09:33 crc kubenswrapper[4492]: I0220 07:09:33.670028 4492 generic.go:334] "Generic (PLEG): container finished" podID="719a407b-3228-4bbc-a600-c08f555ba556" containerID="3dd47efb94381e0eede39e19c59bdb6c288a214c872ccaf5992169e564fb4b5e" exitCode=0 Feb 20 07:09:33 crc kubenswrapper[4492]: I0220 07:09:33.670101 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" event={"ID":"719a407b-3228-4bbc-a600-c08f555ba556","Type":"ContainerDied","Data":"3dd47efb94381e0eede39e19c59bdb6c288a214c872ccaf5992169e564fb4b5e"} Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.017902 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.202343 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-inventory\") pod \"719a407b-3228-4bbc-a600-c08f555ba556\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.202517 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-ssh-key-openstack-edpm-ipam\") pod \"719a407b-3228-4bbc-a600-c08f555ba556\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.202566 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br2lj\" (UniqueName: \"kubernetes.io/projected/719a407b-3228-4bbc-a600-c08f555ba556-kube-api-access-br2lj\") pod \"719a407b-3228-4bbc-a600-c08f555ba556\" (UID: \"719a407b-3228-4bbc-a600-c08f555ba556\") " Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.209436 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/719a407b-3228-4bbc-a600-c08f555ba556-kube-api-access-br2lj" (OuterVolumeSpecName: "kube-api-access-br2lj") pod "719a407b-3228-4bbc-a600-c08f555ba556" (UID: "719a407b-3228-4bbc-a600-c08f555ba556"). InnerVolumeSpecName "kube-api-access-br2lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.226125 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "719a407b-3228-4bbc-a600-c08f555ba556" (UID: "719a407b-3228-4bbc-a600-c08f555ba556"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.226975 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-inventory" (OuterVolumeSpecName: "inventory") pod "719a407b-3228-4bbc-a600-c08f555ba556" (UID: "719a407b-3228-4bbc-a600-c08f555ba556"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.304952 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.304999 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br2lj\" (UniqueName: \"kubernetes.io/projected/719a407b-3228-4bbc-a600-c08f555ba556-kube-api-access-br2lj\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.305013 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/719a407b-3228-4bbc-a600-c08f555ba556-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.695369 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" event={"ID":"719a407b-3228-4bbc-a600-c08f555ba556","Type":"ContainerDied","Data":"be2c9ee9a0593650f89de4c5d04e373cf4f07b9525a457ce208dd1a71ca25a68"} Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.695412 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.695423 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be2c9ee9a0593650f89de4c5d04e373cf4f07b9525a457ce208dd1a71ca25a68" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.869289 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-92sh4"] Feb 20 07:09:35 crc kubenswrapper[4492]: E0220 07:09:35.870386 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719a407b-3228-4bbc-a600-c08f555ba556" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.870419 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="719a407b-3228-4bbc-a600-c08f555ba556" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.870711 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="719a407b-3228-4bbc-a600-c08f555ba556" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.871560 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.873900 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.874184 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.874370 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.874562 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:09:35 crc kubenswrapper[4492]: I0220 07:09:35.881717 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-92sh4"] Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.025707 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c9g7\" (UniqueName: \"kubernetes.io/projected/2505a4cb-335b-4570-9d5a-f0af6666ea9b-kube-api-access-5c9g7\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.025762 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.025801 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.129064 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c9g7\" (UniqueName: \"kubernetes.io/projected/2505a4cb-335b-4570-9d5a-f0af6666ea9b-kube-api-access-5c9g7\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.129136 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.129195 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.134941 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.135320 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.144270 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c9g7\" (UniqueName: \"kubernetes.io/projected/2505a4cb-335b-4570-9d5a-f0af6666ea9b-kube-api-access-5c9g7\") pod \"ssh-known-hosts-edpm-deployment-92sh4\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.195738 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.659357 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-92sh4"] Feb 20 07:09:36 crc kubenswrapper[4492]: I0220 07:09:36.703807 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" event={"ID":"2505a4cb-335b-4570-9d5a-f0af6666ea9b","Type":"ContainerStarted","Data":"2a59a25f3a614fb0022da58dc5d67682a5b65bc271c5fc314c629670ec7b3cf3"} Feb 20 07:09:37 crc kubenswrapper[4492]: I0220 07:09:37.712043 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" event={"ID":"2505a4cb-335b-4570-9d5a-f0af6666ea9b","Type":"ContainerStarted","Data":"07ffaf307e3c8869f358ca737d0d0fdbbac937d127d7b1e0df64a7a5c9d6e076"} Feb 20 07:09:37 crc kubenswrapper[4492]: I0220 07:09:37.735233 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" podStartSLOduration=2.243760015 podStartE2EDuration="2.73520997s" podCreationTimestamp="2026-02-20 07:09:35 +0000 UTC" firstStartedPulling="2026-02-20 07:09:36.660185604 +0000 UTC m=+1733.431474582" lastFinishedPulling="2026-02-20 07:09:37.151635559 +0000 UTC m=+1733.922924537" observedRunningTime="2026-02-20 07:09:37.728297387 +0000 UTC m=+1734.499586365" watchObservedRunningTime="2026-02-20 07:09:37.73520997 +0000 UTC m=+1734.506498948" Feb 20 07:09:42 crc kubenswrapper[4492]: I0220 07:09:42.753005 4492 generic.go:334] "Generic (PLEG): container finished" podID="2505a4cb-335b-4570-9d5a-f0af6666ea9b" containerID="07ffaf307e3c8869f358ca737d0d0fdbbac937d127d7b1e0df64a7a5c9d6e076" exitCode=0 Feb 20 07:09:42 crc kubenswrapper[4492]: I0220 07:09:42.753112 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" event={"ID":"2505a4cb-335b-4570-9d5a-f0af6666ea9b","Type":"ContainerDied","Data":"07ffaf307e3c8869f358ca737d0d0fdbbac937d127d7b1e0df64a7a5c9d6e076"} Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.120391 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.299887 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-ssh-key-openstack-edpm-ipam\") pod \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.299958 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-inventory-0\") pod \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.300003 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c9g7\" (UniqueName: \"kubernetes.io/projected/2505a4cb-335b-4570-9d5a-f0af6666ea9b-kube-api-access-5c9g7\") pod \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\" (UID: \"2505a4cb-335b-4570-9d5a-f0af6666ea9b\") " Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.306332 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2505a4cb-335b-4570-9d5a-f0af6666ea9b-kube-api-access-5c9g7" (OuterVolumeSpecName: "kube-api-access-5c9g7") pod "2505a4cb-335b-4570-9d5a-f0af6666ea9b" (UID: "2505a4cb-335b-4570-9d5a-f0af6666ea9b"). InnerVolumeSpecName "kube-api-access-5c9g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.323539 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2505a4cb-335b-4570-9d5a-f0af6666ea9b" (UID: "2505a4cb-335b-4570-9d5a-f0af6666ea9b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.327859 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "2505a4cb-335b-4570-9d5a-f0af6666ea9b" (UID: "2505a4cb-335b-4570-9d5a-f0af6666ea9b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.403103 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.403137 4492 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2505a4cb-335b-4570-9d5a-f0af6666ea9b-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.403149 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c9g7\" (UniqueName: \"kubernetes.io/projected/2505a4cb-335b-4570-9d5a-f0af6666ea9b-kube-api-access-5c9g7\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.779460 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" event={"ID":"2505a4cb-335b-4570-9d5a-f0af6666ea9b","Type":"ContainerDied","Data":"2a59a25f3a614fb0022da58dc5d67682a5b65bc271c5fc314c629670ec7b3cf3"} Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.779545 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a59a25f3a614fb0022da58dc5d67682a5b65bc271c5fc314c629670ec7b3cf3" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.779637 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-92sh4" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.850231 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh"] Feb 20 07:09:44 crc kubenswrapper[4492]: E0220 07:09:44.850666 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2505a4cb-335b-4570-9d5a-f0af6666ea9b" containerName="ssh-known-hosts-edpm-deployment" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.850685 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505a4cb-335b-4570-9d5a-f0af6666ea9b" containerName="ssh-known-hosts-edpm-deployment" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.850860 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2505a4cb-335b-4570-9d5a-f0af6666ea9b" containerName="ssh-known-hosts-edpm-deployment" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.851570 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.853940 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.854319 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.855904 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.856197 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.866319 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh"] Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.912945 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmw7j\" (UniqueName: \"kubernetes.io/projected/193682fc-046c-4816-a3cf-610626b53e6e-kube-api-access-tmw7j\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.913167 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:44 crc kubenswrapper[4492]: I0220 07:09:44.913239 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.015772 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.015903 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.016177 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmw7j\" (UniqueName: \"kubernetes.io/projected/193682fc-046c-4816-a3cf-610626b53e6e-kube-api-access-tmw7j\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.019692 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.020525 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.032405 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmw7j\" (UniqueName: \"kubernetes.io/projected/193682fc-046c-4816-a3cf-610626b53e6e-kube-api-access-tmw7j\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g2clh\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.167896 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.659631 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh"] Feb 20 07:09:45 crc kubenswrapper[4492]: I0220 07:09:45.792879 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" event={"ID":"193682fc-046c-4816-a3cf-610626b53e6e","Type":"ContainerStarted","Data":"42982586ab6f8965f51f5d895bc97d312f8fa51b000684a5bac872757d7b2f01"} Feb 20 07:09:46 crc kubenswrapper[4492]: I0220 07:09:46.800701 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" event={"ID":"193682fc-046c-4816-a3cf-610626b53e6e","Type":"ContainerStarted","Data":"6efafa04806fe9b24ef810b98ed446ff8b1b3cd84b93753aeab242f47f41d0dd"} Feb 20 07:09:46 crc kubenswrapper[4492]: I0220 07:09:46.825876 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" podStartSLOduration=2.296438403 podStartE2EDuration="2.825861315s" podCreationTimestamp="2026-02-20 07:09:44 +0000 UTC" firstStartedPulling="2026-02-20 07:09:45.666626585 +0000 UTC m=+1742.437915563" lastFinishedPulling="2026-02-20 07:09:46.196049497 +0000 UTC m=+1742.967338475" observedRunningTime="2026-02-20 07:09:46.818026463 +0000 UTC m=+1743.589315440" watchObservedRunningTime="2026-02-20 07:09:46.825861315 +0000 UTC m=+1743.597150292" Feb 20 07:09:47 crc kubenswrapper[4492]: I0220 07:09:47.557431 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:09:47 crc kubenswrapper[4492]: E0220 07:09:47.558512 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:09:52 crc kubenswrapper[4492]: I0220 07:09:52.853344 4492 generic.go:334] "Generic (PLEG): container finished" podID="193682fc-046c-4816-a3cf-610626b53e6e" containerID="6efafa04806fe9b24ef810b98ed446ff8b1b3cd84b93753aeab242f47f41d0dd" exitCode=0 Feb 20 07:09:52 crc kubenswrapper[4492]: I0220 07:09:52.853427 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" event={"ID":"193682fc-046c-4816-a3cf-610626b53e6e","Type":"ContainerDied","Data":"6efafa04806fe9b24ef810b98ed446ff8b1b3cd84b93753aeab242f47f41d0dd"} Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.055050 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-slhpn"] Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.087862 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-slhpn"] Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.277782 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.319084 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmw7j\" (UniqueName: \"kubernetes.io/projected/193682fc-046c-4816-a3cf-610626b53e6e-kube-api-access-tmw7j\") pod \"193682fc-046c-4816-a3cf-610626b53e6e\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.319486 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-inventory\") pod \"193682fc-046c-4816-a3cf-610626b53e6e\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.319882 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-ssh-key-openstack-edpm-ipam\") pod \"193682fc-046c-4816-a3cf-610626b53e6e\" (UID: \"193682fc-046c-4816-a3cf-610626b53e6e\") " Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.327272 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/193682fc-046c-4816-a3cf-610626b53e6e-kube-api-access-tmw7j" (OuterVolumeSpecName: "kube-api-access-tmw7j") pod "193682fc-046c-4816-a3cf-610626b53e6e" (UID: "193682fc-046c-4816-a3cf-610626b53e6e"). InnerVolumeSpecName "kube-api-access-tmw7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.348135 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-inventory" (OuterVolumeSpecName: "inventory") pod "193682fc-046c-4816-a3cf-610626b53e6e" (UID: "193682fc-046c-4816-a3cf-610626b53e6e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.348530 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "193682fc-046c-4816-a3cf-610626b53e6e" (UID: "193682fc-046c-4816-a3cf-610626b53e6e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.424113 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.424146 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmw7j\" (UniqueName: \"kubernetes.io/projected/193682fc-046c-4816-a3cf-610626b53e6e-kube-api-access-tmw7j\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.424156 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/193682fc-046c-4816-a3cf-610626b53e6e-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.878240 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" event={"ID":"193682fc-046c-4816-a3cf-610626b53e6e","Type":"ContainerDied","Data":"42982586ab6f8965f51f5d895bc97d312f8fa51b000684a5bac872757d7b2f01"} Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.878286 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42982586ab6f8965f51f5d895bc97d312f8fa51b000684a5bac872757d7b2f01" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.878315 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g2clh" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.955575 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w"] Feb 20 07:09:54 crc kubenswrapper[4492]: E0220 07:09:54.955995 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="193682fc-046c-4816-a3cf-610626b53e6e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.956026 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="193682fc-046c-4816-a3cf-610626b53e6e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.956288 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="193682fc-046c-4816-a3cf-610626b53e6e" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.957029 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.960771 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.960979 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.961487 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.965656 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w"] Feb 20 07:09:54 crc kubenswrapper[4492]: I0220 07:09:54.969136 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.038504 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.038582 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.038997 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl6m5\" (UniqueName: \"kubernetes.io/projected/b68494be-4b9d-4931-ad0a-5899e5ccf709-kube-api-access-pl6m5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.141113 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.141202 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.141461 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl6m5\" (UniqueName: \"kubernetes.io/projected/b68494be-4b9d-4931-ad0a-5899e5ccf709-kube-api-access-pl6m5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.146585 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.146597 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.159013 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl6m5\" (UniqueName: \"kubernetes.io/projected/b68494be-4b9d-4931-ad0a-5899e5ccf709-kube-api-access-pl6m5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l565w\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.271903 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.567760 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ccbd7a5-60ae-4701-8365-bf880455f53e" path="/var/lib/kubelet/pods/7ccbd7a5-60ae-4701-8365-bf880455f53e/volumes" Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.770994 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w"] Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.785835 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:09:55 crc kubenswrapper[4492]: I0220 07:09:55.888618 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" event={"ID":"b68494be-4b9d-4931-ad0a-5899e5ccf709","Type":"ContainerStarted","Data":"e4c9d2b9e5e0943afcdee3bb43a8a58f8ff56cd747133471e3a412a6fc732d29"} Feb 20 07:09:56 crc kubenswrapper[4492]: I0220 07:09:56.899280 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" event={"ID":"b68494be-4b9d-4931-ad0a-5899e5ccf709","Type":"ContainerStarted","Data":"470665416ef3b8c68cad9b7efb440ba239963f082522114313065bab23416a87"} Feb 20 07:09:56 crc kubenswrapper[4492]: I0220 07:09:56.924326 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" podStartSLOduration=2.407868487 podStartE2EDuration="2.924307518s" podCreationTimestamp="2026-02-20 07:09:54 +0000 UTC" firstStartedPulling="2026-02-20 07:09:55.785616022 +0000 UTC m=+1752.556905000" lastFinishedPulling="2026-02-20 07:09:56.302055062 +0000 UTC m=+1753.073344031" observedRunningTime="2026-02-20 07:09:56.91860665 +0000 UTC m=+1753.689895628" watchObservedRunningTime="2026-02-20 07:09:56.924307518 +0000 UTC m=+1753.695596496" Feb 20 07:10:00 crc kubenswrapper[4492]: I0220 07:10:00.557390 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:10:00 crc kubenswrapper[4492]: E0220 07:10:00.558063 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:10:03 crc kubenswrapper[4492]: I0220 07:10:03.961213 4492 generic.go:334] "Generic (PLEG): container finished" podID="b68494be-4b9d-4931-ad0a-5899e5ccf709" containerID="470665416ef3b8c68cad9b7efb440ba239963f082522114313065bab23416a87" exitCode=0 Feb 20 07:10:03 crc kubenswrapper[4492]: I0220 07:10:03.961416 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" event={"ID":"b68494be-4b9d-4931-ad0a-5899e5ccf709","Type":"ContainerDied","Data":"470665416ef3b8c68cad9b7efb440ba239963f082522114313065bab23416a87"} Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.300455 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.390422 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-ssh-key-openstack-edpm-ipam\") pod \"b68494be-4b9d-4931-ad0a-5899e5ccf709\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.390785 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-inventory\") pod \"b68494be-4b9d-4931-ad0a-5899e5ccf709\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.390972 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl6m5\" (UniqueName: \"kubernetes.io/projected/b68494be-4b9d-4931-ad0a-5899e5ccf709-kube-api-access-pl6m5\") pod \"b68494be-4b9d-4931-ad0a-5899e5ccf709\" (UID: \"b68494be-4b9d-4931-ad0a-5899e5ccf709\") " Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.395643 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b68494be-4b9d-4931-ad0a-5899e5ccf709-kube-api-access-pl6m5" (OuterVolumeSpecName: "kube-api-access-pl6m5") pod "b68494be-4b9d-4931-ad0a-5899e5ccf709" (UID: "b68494be-4b9d-4931-ad0a-5899e5ccf709"). InnerVolumeSpecName "kube-api-access-pl6m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.413410 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "b68494be-4b9d-4931-ad0a-5899e5ccf709" (UID: "b68494be-4b9d-4931-ad0a-5899e5ccf709"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.414377 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-inventory" (OuterVolumeSpecName: "inventory") pod "b68494be-4b9d-4931-ad0a-5899e5ccf709" (UID: "b68494be-4b9d-4931-ad0a-5899e5ccf709"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.494231 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl6m5\" (UniqueName: \"kubernetes.io/projected/b68494be-4b9d-4931-ad0a-5899e5ccf709-kube-api-access-pl6m5\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.494267 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.494279 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b68494be-4b9d-4931-ad0a-5899e5ccf709-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.980393 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" event={"ID":"b68494be-4b9d-4931-ad0a-5899e5ccf709","Type":"ContainerDied","Data":"e4c9d2b9e5e0943afcdee3bb43a8a58f8ff56cd747133471e3a412a6fc732d29"} Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.980776 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4c9d2b9e5e0943afcdee3bb43a8a58f8ff56cd747133471e3a412a6fc732d29" Feb 20 07:10:05 crc kubenswrapper[4492]: I0220 07:10:05.980638 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l565w" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.071933 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj"] Feb 20 07:10:06 crc kubenswrapper[4492]: E0220 07:10:06.072417 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b68494be-4b9d-4931-ad0a-5899e5ccf709" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.072441 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b68494be-4b9d-4931-ad0a-5899e5ccf709" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.072661 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b68494be-4b9d-4931-ad0a-5899e5ccf709" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.073385 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.077051 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.079721 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.080045 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.083390 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.085433 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.085583 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.085753 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.085906 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.087574 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj"] Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110312 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110452 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110492 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110520 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tbpc\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-kube-api-access-4tbpc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110591 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110659 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110682 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110704 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110750 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110789 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110840 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110874 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110939 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.110991 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213134 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213215 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213252 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213358 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213412 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213439 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tbpc\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-kube-api-access-4tbpc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213540 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213620 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213681 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213704 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213767 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213827 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213899 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.213938 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.223353 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.223500 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.224138 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.225207 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.226503 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.229358 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.229525 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.229767 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.230385 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.230510 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.230623 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.233043 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tbpc\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-kube-api-access-4tbpc\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.235715 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.241079 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.392046 4492 scope.go:117] "RemoveContainer" containerID="87e591d8080db489d7bfc1c3018db2736c372426f84b947b74109bffb920aba8" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.409847 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.415748 4492 scope.go:117] "RemoveContainer" containerID="aeaff778beb14841dc9c2652898e241fcc093e71ae51937fee55d8ed24173096" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.564020 4492 scope.go:117] "RemoveContainer" containerID="aa0246f17b0e5dd8261abc5bd0187312e71fe50e6481202bf3406d51922e0faa" Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.947265 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj"] Feb 20 07:10:06 crc kubenswrapper[4492]: I0220 07:10:06.990756 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" event={"ID":"3bcfbb69-f489-46a5-b156-1b17dd76fab5","Type":"ContainerStarted","Data":"485c633c52772712f1e0e2cf0d19038896fec694de31ecc7563e1c27030931fd"} Feb 20 07:10:08 crc kubenswrapper[4492]: I0220 07:10:08.003536 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" event={"ID":"3bcfbb69-f489-46a5-b156-1b17dd76fab5","Type":"ContainerStarted","Data":"d48e0ca23cce75ce6b4ea9de96e3224565eea338e3f6da6e09a92a8e9af5b6b6"} Feb 20 07:10:08 crc kubenswrapper[4492]: I0220 07:10:08.052445 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" podStartSLOduration=1.557976131 podStartE2EDuration="2.052424039s" podCreationTimestamp="2026-02-20 07:10:06 +0000 UTC" firstStartedPulling="2026-02-20 07:10:06.952956586 +0000 UTC m=+1763.724245564" lastFinishedPulling="2026-02-20 07:10:07.447404495 +0000 UTC m=+1764.218693472" observedRunningTime="2026-02-20 07:10:08.028824593 +0000 UTC m=+1764.800113581" watchObservedRunningTime="2026-02-20 07:10:08.052424039 +0000 UTC m=+1764.823713007" Feb 20 07:10:15 crc kubenswrapper[4492]: I0220 07:10:15.557176 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:10:15 crc kubenswrapper[4492]: E0220 07:10:15.557725 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:10:28 crc kubenswrapper[4492]: I0220 07:10:28.556801 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:10:28 crc kubenswrapper[4492]: E0220 07:10:28.557687 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:10:36 crc kubenswrapper[4492]: I0220 07:10:36.236017 4492 generic.go:334] "Generic (PLEG): container finished" podID="3bcfbb69-f489-46a5-b156-1b17dd76fab5" containerID="d48e0ca23cce75ce6b4ea9de96e3224565eea338e3f6da6e09a92a8e9af5b6b6" exitCode=0 Feb 20 07:10:36 crc kubenswrapper[4492]: I0220 07:10:36.236099 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" event={"ID":"3bcfbb69-f489-46a5-b156-1b17dd76fab5","Type":"ContainerDied","Data":"d48e0ca23cce75ce6b4ea9de96e3224565eea338e3f6da6e09a92a8e9af5b6b6"} Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.572995 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595459 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595544 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595582 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ssh-key-openstack-edpm-ipam\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595615 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-inventory\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595728 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-libvirt-combined-ca-bundle\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595754 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-nova-combined-ca-bundle\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595788 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tbpc\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-kube-api-access-4tbpc\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595806 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595828 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-neutron-metadata-combined-ca-bundle\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595849 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-telemetry-combined-ca-bundle\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595889 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-bootstrap-combined-ca-bundle\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595929 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ovn-combined-ca-bundle\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.595978 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.596045 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-repo-setup-combined-ca-bundle\") pod \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\" (UID: \"3bcfbb69-f489-46a5-b156-1b17dd76fab5\") " Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.606667 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.614461 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.614864 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.617164 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.617571 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.618861 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.619716 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-kube-api-access-4tbpc" (OuterVolumeSpecName: "kube-api-access-4tbpc") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "kube-api-access-4tbpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.619880 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.621418 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.621996 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.625755 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.634881 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.640813 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-inventory" (OuterVolumeSpecName: "inventory") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.640836 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3bcfbb69-f489-46a5-b156-1b17dd76fab5" (UID: "3bcfbb69-f489-46a5-b156-1b17dd76fab5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699012 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699050 4492 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699063 4492 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699072 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tbpc\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-kube-api-access-4tbpc\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699081 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699091 4492 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699101 4492 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699112 4492 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699120 4492 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699132 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699143 4492 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699152 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699161 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3bcfbb69-f489-46a5-b156-1b17dd76fab5-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:37 crc kubenswrapper[4492]: I0220 07:10:37.699171 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3bcfbb69-f489-46a5-b156-1b17dd76fab5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.265328 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" event={"ID":"3bcfbb69-f489-46a5-b156-1b17dd76fab5","Type":"ContainerDied","Data":"485c633c52772712f1e0e2cf0d19038896fec694de31ecc7563e1c27030931fd"} Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.265373 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="485c633c52772712f1e0e2cf0d19038896fec694de31ecc7563e1c27030931fd" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.265434 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.375928 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr"] Feb 20 07:10:38 crc kubenswrapper[4492]: E0220 07:10:38.376595 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bcfbb69-f489-46a5-b156-1b17dd76fab5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.376626 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bcfbb69-f489-46a5-b156-1b17dd76fab5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.376948 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bcfbb69-f489-46a5-b156-1b17dd76fab5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.378048 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.380370 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.380544 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.380688 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.381620 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.381747 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.405248 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr"] Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.411129 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.411280 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.411538 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.411628 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.411705 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mhvm\" (UniqueName: \"kubernetes.io/projected/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-kube-api-access-8mhvm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.513562 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mhvm\" (UniqueName: \"kubernetes.io/projected/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-kube-api-access-8mhvm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.513911 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.514005 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.514217 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.514300 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.516066 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.520190 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.529956 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.531226 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.535205 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mhvm\" (UniqueName: \"kubernetes.io/projected/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-kube-api-access-8mhvm\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-75vqr\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:38 crc kubenswrapper[4492]: I0220 07:10:38.713659 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:10:39 crc kubenswrapper[4492]: I0220 07:10:39.172906 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr"] Feb 20 07:10:39 crc kubenswrapper[4492]: W0220 07:10:39.177840 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f5d5f14_0d7b_4117_9651_c1b16f9722f1.slice/crio-2f5655c0976b3fd5eb23ab87ab20faa075671b370e0955cebedfbf003065123b WatchSource:0}: Error finding container 2f5655c0976b3fd5eb23ab87ab20faa075671b370e0955cebedfbf003065123b: Status 404 returned error can't find the container with id 2f5655c0976b3fd5eb23ab87ab20faa075671b370e0955cebedfbf003065123b Feb 20 07:10:39 crc kubenswrapper[4492]: I0220 07:10:39.273182 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" event={"ID":"1f5d5f14-0d7b-4117-9651-c1b16f9722f1","Type":"ContainerStarted","Data":"2f5655c0976b3fd5eb23ab87ab20faa075671b370e0955cebedfbf003065123b"} Feb 20 07:10:40 crc kubenswrapper[4492]: I0220 07:10:40.281722 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" event={"ID":"1f5d5f14-0d7b-4117-9651-c1b16f9722f1","Type":"ContainerStarted","Data":"0e73bc6a1b125a0091ab550a5b722d96537d3daee196cf8df7f59cda43c1e9e5"} Feb 20 07:10:40 crc kubenswrapper[4492]: I0220 07:10:40.299517 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" podStartSLOduration=1.802345929 podStartE2EDuration="2.299504418s" podCreationTimestamp="2026-02-20 07:10:38 +0000 UTC" firstStartedPulling="2026-02-20 07:10:39.180991368 +0000 UTC m=+1795.952280345" lastFinishedPulling="2026-02-20 07:10:39.678149856 +0000 UTC m=+1796.449438834" observedRunningTime="2026-02-20 07:10:40.295943484 +0000 UTC m=+1797.067232462" watchObservedRunningTime="2026-02-20 07:10:40.299504418 +0000 UTC m=+1797.070793386" Feb 20 07:10:40 crc kubenswrapper[4492]: I0220 07:10:40.557177 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:10:40 crc kubenswrapper[4492]: E0220 07:10:40.557994 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:10:53 crc kubenswrapper[4492]: I0220 07:10:53.562820 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:10:53 crc kubenswrapper[4492]: E0220 07:10:53.563736 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:11:08 crc kubenswrapper[4492]: I0220 07:11:08.557669 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:11:08 crc kubenswrapper[4492]: E0220 07:11:08.558385 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:11:20 crc kubenswrapper[4492]: I0220 07:11:20.557585 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:11:21 crc kubenswrapper[4492]: I0220 07:11:21.596704 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"766c6946637b4404c8c01854dc3221d37256795a37babb77f668a65574b03c74"} Feb 20 07:11:30 crc kubenswrapper[4492]: I0220 07:11:30.675148 4492 generic.go:334] "Generic (PLEG): container finished" podID="1f5d5f14-0d7b-4117-9651-c1b16f9722f1" containerID="0e73bc6a1b125a0091ab550a5b722d96537d3daee196cf8df7f59cda43c1e9e5" exitCode=0 Feb 20 07:11:30 crc kubenswrapper[4492]: I0220 07:11:30.675228 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" event={"ID":"1f5d5f14-0d7b-4117-9651-c1b16f9722f1","Type":"ContainerDied","Data":"0e73bc6a1b125a0091ab550a5b722d96537d3daee196cf8df7f59cda43c1e9e5"} Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.042664 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.210642 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-inventory\") pod \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.210750 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ssh-key-openstack-edpm-ipam\") pod \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.210773 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovn-combined-ca-bundle\") pod \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.210793 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mhvm\" (UniqueName: \"kubernetes.io/projected/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-kube-api-access-8mhvm\") pod \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.210906 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovncontroller-config-0\") pod \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\" (UID: \"1f5d5f14-0d7b-4117-9651-c1b16f9722f1\") " Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.217179 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "1f5d5f14-0d7b-4117-9651-c1b16f9722f1" (UID: "1f5d5f14-0d7b-4117-9651-c1b16f9722f1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.218160 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-kube-api-access-8mhvm" (OuterVolumeSpecName: "kube-api-access-8mhvm") pod "1f5d5f14-0d7b-4117-9651-c1b16f9722f1" (UID: "1f5d5f14-0d7b-4117-9651-c1b16f9722f1"). InnerVolumeSpecName "kube-api-access-8mhvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.233653 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "1f5d5f14-0d7b-4117-9651-c1b16f9722f1" (UID: "1f5d5f14-0d7b-4117-9651-c1b16f9722f1"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.235906 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-inventory" (OuterVolumeSpecName: "inventory") pod "1f5d5f14-0d7b-4117-9651-c1b16f9722f1" (UID: "1f5d5f14-0d7b-4117-9651-c1b16f9722f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.239157 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1f5d5f14-0d7b-4117-9651-c1b16f9722f1" (UID: "1f5d5f14-0d7b-4117-9651-c1b16f9722f1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.313792 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.313880 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.314021 4492 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.314279 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mhvm\" (UniqueName: \"kubernetes.io/projected/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-kube-api-access-8mhvm\") on node \"crc\" DevicePath \"\"" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.314335 4492 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/1f5d5f14-0d7b-4117-9651-c1b16f9722f1-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.696872 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" event={"ID":"1f5d5f14-0d7b-4117-9651-c1b16f9722f1","Type":"ContainerDied","Data":"2f5655c0976b3fd5eb23ab87ab20faa075671b370e0955cebedfbf003065123b"} Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.697067 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f5655c0976b3fd5eb23ab87ab20faa075671b370e0955cebedfbf003065123b" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.696988 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-75vqr" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.860351 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs"] Feb 20 07:11:32 crc kubenswrapper[4492]: E0220 07:11:32.861055 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f5d5f14-0d7b-4117-9651-c1b16f9722f1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.861156 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f5d5f14-0d7b-4117-9651-c1b16f9722f1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.861382 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f5d5f14-0d7b-4117-9651-c1b16f9722f1" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.862228 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.864770 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.865198 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.865535 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.865729 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.865916 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.866036 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.869783 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs"] Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.926164 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.926239 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.926272 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.926371 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.926620 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:32 crc kubenswrapper[4492]: I0220 07:11:32.926870 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrdh5\" (UniqueName: \"kubernetes.io/projected/e2680c81-598b-42bf-976e-2c595d48a5b7-kube-api-access-xrdh5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.037342 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.037418 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.037485 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrdh5\" (UniqueName: \"kubernetes.io/projected/e2680c81-598b-42bf-976e-2c595d48a5b7-kube-api-access-xrdh5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.037539 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.037565 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.037581 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.048028 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.053574 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.054335 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.060928 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.075851 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.085043 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrdh5\" (UniqueName: \"kubernetes.io/projected/e2680c81-598b-42bf-976e-2c595d48a5b7-kube-api-access-xrdh5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.181171 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:11:33 crc kubenswrapper[4492]: I0220 07:11:33.696977 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs"] Feb 20 07:11:34 crc kubenswrapper[4492]: I0220 07:11:34.715246 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" event={"ID":"e2680c81-598b-42bf-976e-2c595d48a5b7","Type":"ContainerStarted","Data":"e7fa2abddf04ff84473cd64427db4668d256158879027dfd2387dda8c1e856f0"} Feb 20 07:11:34 crc kubenswrapper[4492]: I0220 07:11:34.715825 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" event={"ID":"e2680c81-598b-42bf-976e-2c595d48a5b7","Type":"ContainerStarted","Data":"53b9f01c9c02d83a052797747de6cdbce6f5ee6340ff123c15e1307f44bf5db2"} Feb 20 07:11:34 crc kubenswrapper[4492]: I0220 07:11:34.743038 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" podStartSLOduration=2.254860072 podStartE2EDuration="2.743026347s" podCreationTimestamp="2026-02-20 07:11:32 +0000 UTC" firstStartedPulling="2026-02-20 07:11:33.706225359 +0000 UTC m=+1850.477514336" lastFinishedPulling="2026-02-20 07:11:34.194391633 +0000 UTC m=+1850.965680611" observedRunningTime="2026-02-20 07:11:34.734720648 +0000 UTC m=+1851.506009626" watchObservedRunningTime="2026-02-20 07:11:34.743026347 +0000 UTC m=+1851.514315325" Feb 20 07:12:13 crc kubenswrapper[4492]: I0220 07:12:13.042098 4492 generic.go:334] "Generic (PLEG): container finished" podID="e2680c81-598b-42bf-976e-2c595d48a5b7" containerID="e7fa2abddf04ff84473cd64427db4668d256158879027dfd2387dda8c1e856f0" exitCode=0 Feb 20 07:12:13 crc kubenswrapper[4492]: I0220 07:12:13.042270 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" event={"ID":"e2680c81-598b-42bf-976e-2c595d48a5b7","Type":"ContainerDied","Data":"e7fa2abddf04ff84473cd64427db4668d256158879027dfd2387dda8c1e856f0"} Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.412568 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.439012 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-inventory\") pod \"e2680c81-598b-42bf-976e-2c595d48a5b7\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.439059 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrdh5\" (UniqueName: \"kubernetes.io/projected/e2680c81-598b-42bf-976e-2c595d48a5b7-kube-api-access-xrdh5\") pod \"e2680c81-598b-42bf-976e-2c595d48a5b7\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.439112 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-ssh-key-openstack-edpm-ipam\") pod \"e2680c81-598b-42bf-976e-2c595d48a5b7\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.439150 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-metadata-combined-ca-bundle\") pod \"e2680c81-598b-42bf-976e-2c595d48a5b7\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.439219 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-ovn-metadata-agent-neutron-config-0\") pod \"e2680c81-598b-42bf-976e-2c595d48a5b7\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.439354 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-nova-metadata-neutron-config-0\") pod \"e2680c81-598b-42bf-976e-2c595d48a5b7\" (UID: \"e2680c81-598b-42bf-976e-2c595d48a5b7\") " Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.452771 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e2680c81-598b-42bf-976e-2c595d48a5b7" (UID: "e2680c81-598b-42bf-976e-2c595d48a5b7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.461910 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2680c81-598b-42bf-976e-2c595d48a5b7-kube-api-access-xrdh5" (OuterVolumeSpecName: "kube-api-access-xrdh5") pod "e2680c81-598b-42bf-976e-2c595d48a5b7" (UID: "e2680c81-598b-42bf-976e-2c595d48a5b7"). InnerVolumeSpecName "kube-api-access-xrdh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.465676 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-inventory" (OuterVolumeSpecName: "inventory") pod "e2680c81-598b-42bf-976e-2c595d48a5b7" (UID: "e2680c81-598b-42bf-976e-2c595d48a5b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.468589 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "e2680c81-598b-42bf-976e-2c595d48a5b7" (UID: "e2680c81-598b-42bf-976e-2c595d48a5b7"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.468678 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "e2680c81-598b-42bf-976e-2c595d48a5b7" (UID: "e2680c81-598b-42bf-976e-2c595d48a5b7"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.469618 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e2680c81-598b-42bf-976e-2c595d48a5b7" (UID: "e2680c81-598b-42bf-976e-2c595d48a5b7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.543105 4492 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.543141 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.543153 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrdh5\" (UniqueName: \"kubernetes.io/projected/e2680c81-598b-42bf-976e-2c595d48a5b7-kube-api-access-xrdh5\") on node \"crc\" DevicePath \"\"" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.543165 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.543178 4492 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:12:14 crc kubenswrapper[4492]: I0220 07:12:14.543187 4492 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e2680c81-598b-42bf-976e-2c595d48a5b7-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.059887 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" event={"ID":"e2680c81-598b-42bf-976e-2c595d48a5b7","Type":"ContainerDied","Data":"53b9f01c9c02d83a052797747de6cdbce6f5ee6340ff123c15e1307f44bf5db2"} Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.060290 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53b9f01c9c02d83a052797747de6cdbce6f5ee6340ff123c15e1307f44bf5db2" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.059940 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.154376 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6"] Feb 20 07:12:15 crc kubenswrapper[4492]: E0220 07:12:15.154914 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2680c81-598b-42bf-976e-2c595d48a5b7" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.154991 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2680c81-598b-42bf-976e-2c595d48a5b7" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.155233 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2680c81-598b-42bf-976e-2c595d48a5b7" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.155955 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.159168 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.161367 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.167724 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.167819 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.167904 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.172457 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6"] Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.256805 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.256873 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.256938 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hfr8\" (UniqueName: \"kubernetes.io/projected/94e8423b-3881-4585-8f99-ef80faebae3c-kube-api-access-2hfr8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.257002 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.257059 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.359263 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.359334 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.359537 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.359619 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.359677 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hfr8\" (UniqueName: \"kubernetes.io/projected/94e8423b-3881-4585-8f99-ef80faebae3c-kube-api-access-2hfr8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.366174 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.371048 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.372198 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.372317 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.374615 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hfr8\" (UniqueName: \"kubernetes.io/projected/94e8423b-3881-4585-8f99-ef80faebae3c-kube-api-access-2hfr8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.471932 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:12:15 crc kubenswrapper[4492]: I0220 07:12:15.956220 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6"] Feb 20 07:12:16 crc kubenswrapper[4492]: I0220 07:12:16.071693 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" event={"ID":"94e8423b-3881-4585-8f99-ef80faebae3c","Type":"ContainerStarted","Data":"abf426e2015b9eae387dd5ef4a88238d8d3c341bc673388d1b3923629519e7c4"} Feb 20 07:12:17 crc kubenswrapper[4492]: I0220 07:12:17.085445 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" event={"ID":"94e8423b-3881-4585-8f99-ef80faebae3c","Type":"ContainerStarted","Data":"1899dc481a05451b0d6a2ce620097724fafee4f766246bfd166adcb370f3da90"} Feb 20 07:12:17 crc kubenswrapper[4492]: I0220 07:12:17.111404 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" podStartSLOduration=1.5618528729999999 podStartE2EDuration="2.111382715s" podCreationTimestamp="2026-02-20 07:12:15 +0000 UTC" firstStartedPulling="2026-02-20 07:12:15.96160653 +0000 UTC m=+1892.732895507" lastFinishedPulling="2026-02-20 07:12:16.511136371 +0000 UTC m=+1893.282425349" observedRunningTime="2026-02-20 07:12:17.097885768 +0000 UTC m=+1893.869174746" watchObservedRunningTime="2026-02-20 07:12:17.111382715 +0000 UTC m=+1893.882671693" Feb 20 07:13:39 crc kubenswrapper[4492]: I0220 07:13:39.311108 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:13:39 crc kubenswrapper[4492]: I0220 07:13:39.311505 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:14:09 crc kubenswrapper[4492]: I0220 07:14:09.311013 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:14:09 crc kubenswrapper[4492]: I0220 07:14:09.312067 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:14:39 crc kubenswrapper[4492]: I0220 07:14:39.310876 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:14:39 crc kubenswrapper[4492]: I0220 07:14:39.311389 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:14:39 crc kubenswrapper[4492]: I0220 07:14:39.311435 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:14:39 crc kubenswrapper[4492]: I0220 07:14:39.312283 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"766c6946637b4404c8c01854dc3221d37256795a37babb77f668a65574b03c74"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:14:39 crc kubenswrapper[4492]: I0220 07:14:39.312335 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://766c6946637b4404c8c01854dc3221d37256795a37babb77f668a65574b03c74" gracePeriod=600 Feb 20 07:14:40 crc kubenswrapper[4492]: I0220 07:14:40.047894 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="766c6946637b4404c8c01854dc3221d37256795a37babb77f668a65574b03c74" exitCode=0 Feb 20 07:14:40 crc kubenswrapper[4492]: I0220 07:14:40.047972 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"766c6946637b4404c8c01854dc3221d37256795a37babb77f668a65574b03c74"} Feb 20 07:14:40 crc kubenswrapper[4492]: I0220 07:14:40.048407 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97"} Feb 20 07:14:40 crc kubenswrapper[4492]: I0220 07:14:40.048433 4492 scope.go:117] "RemoveContainer" containerID="ae964431b36ac88573c19527e594d7c536ffe697a419df4c36d5bbc5477ebafe" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.147067 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv"] Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.149275 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.152215 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.152410 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.155187 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv"] Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.238281 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-secret-volume\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.238662 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x9x5\" (UniqueName: \"kubernetes.io/projected/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-kube-api-access-5x9x5\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.238740 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-config-volume\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.340527 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-secret-volume\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.340719 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x9x5\" (UniqueName: \"kubernetes.io/projected/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-kube-api-access-5x9x5\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.340843 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-config-volume\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.341813 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-config-volume\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.350637 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-secret-volume\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.357647 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x9x5\" (UniqueName: \"kubernetes.io/projected/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-kube-api-access-5x9x5\") pod \"collect-profiles-29526195-h6glv\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.484355 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:00 crc kubenswrapper[4492]: I0220 07:15:00.902282 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv"] Feb 20 07:15:01 crc kubenswrapper[4492]: I0220 07:15:01.217763 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" event={"ID":"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8","Type":"ContainerStarted","Data":"70214d122c4fe249cd6cbaf8e35355228ebdc9ad69ebe5eb885d5f0ea56382d7"} Feb 20 07:15:01 crc kubenswrapper[4492]: I0220 07:15:01.218176 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" event={"ID":"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8","Type":"ContainerStarted","Data":"79bdb11f756cbc4f3f25581a62a98b0c7b0079b0e12e6d08b9fd4a84be885e88"} Feb 20 07:15:01 crc kubenswrapper[4492]: I0220 07:15:01.235567 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" podStartSLOduration=1.235544725 podStartE2EDuration="1.235544725s" podCreationTimestamp="2026-02-20 07:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:15:01.230544558 +0000 UTC m=+2058.001833535" watchObservedRunningTime="2026-02-20 07:15:01.235544725 +0000 UTC m=+2058.006833693" Feb 20 07:15:02 crc kubenswrapper[4492]: I0220 07:15:02.228262 4492 generic.go:334] "Generic (PLEG): container finished" podID="be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" containerID="70214d122c4fe249cd6cbaf8e35355228ebdc9ad69ebe5eb885d5f0ea56382d7" exitCode=0 Feb 20 07:15:02 crc kubenswrapper[4492]: I0220 07:15:02.228407 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" event={"ID":"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8","Type":"ContainerDied","Data":"70214d122c4fe249cd6cbaf8e35355228ebdc9ad69ebe5eb885d5f0ea56382d7"} Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.472992 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.621768 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-secret-volume\") pod \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.621880 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-config-volume\") pod \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.621960 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5x9x5\" (UniqueName: \"kubernetes.io/projected/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-kube-api-access-5x9x5\") pod \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\" (UID: \"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8\") " Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.622747 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-config-volume" (OuterVolumeSpecName: "config-volume") pod "be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" (UID: "be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.627231 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" (UID: "be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.627654 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-kube-api-access-5x9x5" (OuterVolumeSpecName: "kube-api-access-5x9x5") pod "be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" (UID: "be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8"). InnerVolumeSpecName "kube-api-access-5x9x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.724399 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5x9x5\" (UniqueName: \"kubernetes.io/projected/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-kube-api-access-5x9x5\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.724425 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:03 crc kubenswrapper[4492]: I0220 07:15:03.724435 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:04 crc kubenswrapper[4492]: I0220 07:15:04.243859 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" event={"ID":"be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8","Type":"ContainerDied","Data":"79bdb11f756cbc4f3f25581a62a98b0c7b0079b0e12e6d08b9fd4a84be885e88"} Feb 20 07:15:04 crc kubenswrapper[4492]: I0220 07:15:04.244113 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79bdb11f756cbc4f3f25581a62a98b0c7b0079b0e12e6d08b9fd4a84be885e88" Feb 20 07:15:04 crc kubenswrapper[4492]: I0220 07:15:04.243887 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv" Feb 20 07:15:04 crc kubenswrapper[4492]: I0220 07:15:04.298947 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4"] Feb 20 07:15:04 crc kubenswrapper[4492]: I0220 07:15:04.305236 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526150-tswn4"] Feb 20 07:15:05 crc kubenswrapper[4492]: I0220 07:15:05.566299 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98478738-51a1-4daf-b1ca-34c21711fb40" path="/var/lib/kubelet/pods/98478738-51a1-4daf-b1ca-34c21711fb40/volumes" Feb 20 07:15:06 crc kubenswrapper[4492]: I0220 07:15:06.720834 4492 scope.go:117] "RemoveContainer" containerID="375b3d46d8a7f93aae5dd415c252610d14e35734e3c403d6bd9ea4011919316b" Feb 20 07:15:14 crc kubenswrapper[4492]: I0220 07:15:14.319611 4492 generic.go:334] "Generic (PLEG): container finished" podID="94e8423b-3881-4585-8f99-ef80faebae3c" containerID="1899dc481a05451b0d6a2ce620097724fafee4f766246bfd166adcb370f3da90" exitCode=0 Feb 20 07:15:14 crc kubenswrapper[4492]: I0220 07:15:14.319722 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" event={"ID":"94e8423b-3881-4585-8f99-ef80faebae3c","Type":"ContainerDied","Data":"1899dc481a05451b0d6a2ce620097724fafee4f766246bfd166adcb370f3da90"} Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.829449 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.978458 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hfr8\" (UniqueName: \"kubernetes.io/projected/94e8423b-3881-4585-8f99-ef80faebae3c-kube-api-access-2hfr8\") pod \"94e8423b-3881-4585-8f99-ef80faebae3c\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.978669 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-ssh-key-openstack-edpm-ipam\") pod \"94e8423b-3881-4585-8f99-ef80faebae3c\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.978792 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-inventory\") pod \"94e8423b-3881-4585-8f99-ef80faebae3c\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.979426 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-secret-0\") pod \"94e8423b-3881-4585-8f99-ef80faebae3c\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.979613 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-combined-ca-bundle\") pod \"94e8423b-3881-4585-8f99-ef80faebae3c\" (UID: \"94e8423b-3881-4585-8f99-ef80faebae3c\") " Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.985250 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "94e8423b-3881-4585-8f99-ef80faebae3c" (UID: "94e8423b-3881-4585-8f99-ef80faebae3c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:15:15 crc kubenswrapper[4492]: I0220 07:15:15.985886 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94e8423b-3881-4585-8f99-ef80faebae3c-kube-api-access-2hfr8" (OuterVolumeSpecName: "kube-api-access-2hfr8") pod "94e8423b-3881-4585-8f99-ef80faebae3c" (UID: "94e8423b-3881-4585-8f99-ef80faebae3c"). InnerVolumeSpecName "kube-api-access-2hfr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.006692 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "94e8423b-3881-4585-8f99-ef80faebae3c" (UID: "94e8423b-3881-4585-8f99-ef80faebae3c"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.006663 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "94e8423b-3881-4585-8f99-ef80faebae3c" (UID: "94e8423b-3881-4585-8f99-ef80faebae3c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.007291 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-inventory" (OuterVolumeSpecName: "inventory") pod "94e8423b-3881-4585-8f99-ef80faebae3c" (UID: "94e8423b-3881-4585-8f99-ef80faebae3c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.083013 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.083046 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.083060 4492 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.083069 4492 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e8423b-3881-4585-8f99-ef80faebae3c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.083082 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hfr8\" (UniqueName: \"kubernetes.io/projected/94e8423b-3881-4585-8f99-ef80faebae3c-kube-api-access-2hfr8\") on node \"crc\" DevicePath \"\"" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.339069 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" event={"ID":"94e8423b-3881-4585-8f99-ef80faebae3c","Type":"ContainerDied","Data":"abf426e2015b9eae387dd5ef4a88238d8d3c341bc673388d1b3923629519e7c4"} Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.339407 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abf426e2015b9eae387dd5ef4a88238d8d3c341bc673388d1b3923629519e7c4" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.339505 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.432866 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2"] Feb 20 07:15:16 crc kubenswrapper[4492]: E0220 07:15:16.434120 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" containerName="collect-profiles" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.434142 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" containerName="collect-profiles" Feb 20 07:15:16 crc kubenswrapper[4492]: E0220 07:15:16.434185 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94e8423b-3881-4585-8f99-ef80faebae3c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.434194 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="94e8423b-3881-4585-8f99-ef80faebae3c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.435145 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="94e8423b-3881-4585-8f99-ef80faebae3c" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.435175 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" containerName="collect-profiles" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.437937 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.447210 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.447596 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.448194 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.448497 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.448600 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.455215 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.460126 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.489147 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2"] Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607260 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607331 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8jvj\" (UniqueName: \"kubernetes.io/projected/51cf4d1e-7eb2-4882-bd28-c34e75100a67-kube-api-access-f8jvj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607466 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607577 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607662 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607718 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607895 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.607964 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.608091 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.608128 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.608147 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.710453 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.710566 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.710628 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711353 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711402 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711553 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711584 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711607 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711655 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711678 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8jvj\" (UniqueName: \"kubernetes.io/projected/51cf4d1e-7eb2-4882-bd28-c34e75100a67-kube-api-access-f8jvj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.711718 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.712461 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.717346 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.717654 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.718074 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.722229 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.722303 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.722541 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.722618 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.722672 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.724143 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.729375 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8jvj\" (UniqueName: \"kubernetes.io/projected/51cf4d1e-7eb2-4882-bd28-c34e75100a67-kube-api-access-f8jvj\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jqcw2\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:16 crc kubenswrapper[4492]: I0220 07:15:16.767761 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:15:17 crc kubenswrapper[4492]: I0220 07:15:17.249149 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2"] Feb 20 07:15:17 crc kubenswrapper[4492]: I0220 07:15:17.258932 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:15:17 crc kubenswrapper[4492]: I0220 07:15:17.348966 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" event={"ID":"51cf4d1e-7eb2-4882-bd28-c34e75100a67","Type":"ContainerStarted","Data":"4192c14c9f5eaf7e52a42b5acd0dec75d1c5f5bf1f65b42d30b7d758d86cac87"} Feb 20 07:15:18 crc kubenswrapper[4492]: I0220 07:15:18.360596 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" event={"ID":"51cf4d1e-7eb2-4882-bd28-c34e75100a67","Type":"ContainerStarted","Data":"47994f5167a3a6d97e6ff38c5b0b3cf0a8e8d7306a5d51866a3031b058972d70"} Feb 20 07:15:18 crc kubenswrapper[4492]: I0220 07:15:18.389168 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" podStartSLOduration=1.8672366070000002 podStartE2EDuration="2.389147338s" podCreationTimestamp="2026-02-20 07:15:16 +0000 UTC" firstStartedPulling="2026-02-20 07:15:17.258666238 +0000 UTC m=+2074.029955216" lastFinishedPulling="2026-02-20 07:15:17.780576969 +0000 UTC m=+2074.551865947" observedRunningTime="2026-02-20 07:15:18.380426575 +0000 UTC m=+2075.151715563" watchObservedRunningTime="2026-02-20 07:15:18.389147338 +0000 UTC m=+2075.160436317" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.050871 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gsm2x"] Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.064910 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsm2x"] Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.065050 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.073149 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-catalog-content\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.073210 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55v6g\" (UniqueName: \"kubernetes.io/projected/df9394ea-402a-44ef-b1f6-365eef6d5f6a-kube-api-access-55v6g\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.073315 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-utilities\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.176021 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-utilities\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.176513 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-catalog-content\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.176653 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55v6g\" (UniqueName: \"kubernetes.io/projected/df9394ea-402a-44ef-b1f6-365eef6d5f6a-kube-api-access-55v6g\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.176982 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-catalog-content\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.176982 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-utilities\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.204910 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55v6g\" (UniqueName: \"kubernetes.io/projected/df9394ea-402a-44ef-b1f6-365eef6d5f6a-kube-api-access-55v6g\") pod \"certified-operators-gsm2x\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.384185 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:38 crc kubenswrapper[4492]: I0220 07:16:38.907545 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gsm2x"] Feb 20 07:16:39 crc kubenswrapper[4492]: I0220 07:16:39.006824 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsm2x" event={"ID":"df9394ea-402a-44ef-b1f6-365eef6d5f6a","Type":"ContainerStarted","Data":"e575a709488183cc82105190778e12a17de73aea43bde0ad082f728cb09de4d4"} Feb 20 07:16:39 crc kubenswrapper[4492]: I0220 07:16:39.311320 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:16:39 crc kubenswrapper[4492]: I0220 07:16:39.311498 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:16:40 crc kubenswrapper[4492]: I0220 07:16:40.018521 4492 generic.go:334] "Generic (PLEG): container finished" podID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerID="857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267" exitCode=0 Feb 20 07:16:40 crc kubenswrapper[4492]: I0220 07:16:40.018592 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsm2x" event={"ID":"df9394ea-402a-44ef-b1f6-365eef6d5f6a","Type":"ContainerDied","Data":"857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267"} Feb 20 07:16:41 crc kubenswrapper[4492]: I0220 07:16:41.030626 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsm2x" event={"ID":"df9394ea-402a-44ef-b1f6-365eef6d5f6a","Type":"ContainerStarted","Data":"292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d"} Feb 20 07:16:42 crc kubenswrapper[4492]: I0220 07:16:42.044036 4492 generic.go:334] "Generic (PLEG): container finished" podID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerID="292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d" exitCode=0 Feb 20 07:16:42 crc kubenswrapper[4492]: I0220 07:16:42.044303 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsm2x" event={"ID":"df9394ea-402a-44ef-b1f6-365eef6d5f6a","Type":"ContainerDied","Data":"292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d"} Feb 20 07:16:43 crc kubenswrapper[4492]: I0220 07:16:43.058657 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsm2x" event={"ID":"df9394ea-402a-44ef-b1f6-365eef6d5f6a","Type":"ContainerStarted","Data":"611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2"} Feb 20 07:16:43 crc kubenswrapper[4492]: I0220 07:16:43.077715 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gsm2x" podStartSLOduration=2.546299353 podStartE2EDuration="5.077702544s" podCreationTimestamp="2026-02-20 07:16:38 +0000 UTC" firstStartedPulling="2026-02-20 07:16:40.020620567 +0000 UTC m=+2156.791909535" lastFinishedPulling="2026-02-20 07:16:42.552023748 +0000 UTC m=+2159.323312726" observedRunningTime="2026-02-20 07:16:43.072615192 +0000 UTC m=+2159.843904169" watchObservedRunningTime="2026-02-20 07:16:43.077702544 +0000 UTC m=+2159.848991522" Feb 20 07:16:48 crc kubenswrapper[4492]: I0220 07:16:48.384904 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:48 crc kubenswrapper[4492]: I0220 07:16:48.385785 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:48 crc kubenswrapper[4492]: I0220 07:16:48.432608 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:49 crc kubenswrapper[4492]: I0220 07:16:49.152626 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:49 crc kubenswrapper[4492]: I0220 07:16:49.203181 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gsm2x"] Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.135456 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gsm2x" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="registry-server" containerID="cri-o://611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2" gracePeriod=2 Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.555589 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.707938 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55v6g\" (UniqueName: \"kubernetes.io/projected/df9394ea-402a-44ef-b1f6-365eef6d5f6a-kube-api-access-55v6g\") pod \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.708213 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-utilities\") pod \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.708270 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-catalog-content\") pod \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\" (UID: \"df9394ea-402a-44ef-b1f6-365eef6d5f6a\") " Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.708775 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-utilities" (OuterVolumeSpecName: "utilities") pod "df9394ea-402a-44ef-b1f6-365eef6d5f6a" (UID: "df9394ea-402a-44ef-b1f6-365eef6d5f6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.709179 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.716618 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df9394ea-402a-44ef-b1f6-365eef6d5f6a-kube-api-access-55v6g" (OuterVolumeSpecName: "kube-api-access-55v6g") pod "df9394ea-402a-44ef-b1f6-365eef6d5f6a" (UID: "df9394ea-402a-44ef-b1f6-365eef6d5f6a"). InnerVolumeSpecName "kube-api-access-55v6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.757654 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df9394ea-402a-44ef-b1f6-365eef6d5f6a" (UID: "df9394ea-402a-44ef-b1f6-365eef6d5f6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.811067 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55v6g\" (UniqueName: \"kubernetes.io/projected/df9394ea-402a-44ef-b1f6-365eef6d5f6a-kube-api-access-55v6g\") on node \"crc\" DevicePath \"\"" Feb 20 07:16:51 crc kubenswrapper[4492]: I0220 07:16:51.811103 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df9394ea-402a-44ef-b1f6-365eef6d5f6a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.144807 4492 generic.go:334] "Generic (PLEG): container finished" podID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerID="611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2" exitCode=0 Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.144864 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsm2x" event={"ID":"df9394ea-402a-44ef-b1f6-365eef6d5f6a","Type":"ContainerDied","Data":"611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2"} Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.144901 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gsm2x" event={"ID":"df9394ea-402a-44ef-b1f6-365eef6d5f6a","Type":"ContainerDied","Data":"e575a709488183cc82105190778e12a17de73aea43bde0ad082f728cb09de4d4"} Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.144921 4492 scope.go:117] "RemoveContainer" containerID="611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.145101 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gsm2x" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.162732 4492 scope.go:117] "RemoveContainer" containerID="292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.200263 4492 scope.go:117] "RemoveContainer" containerID="857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.200659 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gsm2x"] Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.211008 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gsm2x"] Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.226485 4492 scope.go:117] "RemoveContainer" containerID="611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2" Feb 20 07:16:52 crc kubenswrapper[4492]: E0220 07:16:52.226925 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2\": container with ID starting with 611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2 not found: ID does not exist" containerID="611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.226961 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2"} err="failed to get container status \"611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2\": rpc error: code = NotFound desc = could not find container \"611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2\": container with ID starting with 611439e2aad6ff7aa3b36223e4a393a3783279649f25442bef7f03a84684c3e2 not found: ID does not exist" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.226986 4492 scope.go:117] "RemoveContainer" containerID="292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d" Feb 20 07:16:52 crc kubenswrapper[4492]: E0220 07:16:52.227176 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d\": container with ID starting with 292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d not found: ID does not exist" containerID="292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.227198 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d"} err="failed to get container status \"292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d\": rpc error: code = NotFound desc = could not find container \"292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d\": container with ID starting with 292930284cb0edceefe21fec36cce47ec4f323cd6b6b11b469e410411b53a20d not found: ID does not exist" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.227214 4492 scope.go:117] "RemoveContainer" containerID="857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267" Feb 20 07:16:52 crc kubenswrapper[4492]: E0220 07:16:52.227398 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267\": container with ID starting with 857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267 not found: ID does not exist" containerID="857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267" Feb 20 07:16:52 crc kubenswrapper[4492]: I0220 07:16:52.227422 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267"} err="failed to get container status \"857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267\": rpc error: code = NotFound desc = could not find container \"857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267\": container with ID starting with 857af6959b0ff24a01b762bb34a56def9b050bb8ddfd97cf73ee03b761cbf267 not found: ID does not exist" Feb 20 07:16:53 crc kubenswrapper[4492]: I0220 07:16:53.566728 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" path="/var/lib/kubelet/pods/df9394ea-402a-44ef-b1f6-365eef6d5f6a/volumes" Feb 20 07:17:09 crc kubenswrapper[4492]: I0220 07:17:09.311366 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:17:09 crc kubenswrapper[4492]: I0220 07:17:09.311787 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:17:12 crc kubenswrapper[4492]: I0220 07:17:12.292774 4492 generic.go:334] "Generic (PLEG): container finished" podID="51cf4d1e-7eb2-4882-bd28-c34e75100a67" containerID="47994f5167a3a6d97e6ff38c5b0b3cf0a8e8d7306a5d51866a3031b058972d70" exitCode=0 Feb 20 07:17:12 crc kubenswrapper[4492]: I0220 07:17:12.292866 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" event={"ID":"51cf4d1e-7eb2-4882-bd28-c34e75100a67","Type":"ContainerDied","Data":"47994f5167a3a6d97e6ff38c5b0b3cf0a8e8d7306a5d51866a3031b058972d70"} Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.656253 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.707068 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-1\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.707345 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-3\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.707379 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8jvj\" (UniqueName: \"kubernetes.io/projected/51cf4d1e-7eb2-4882-bd28-c34e75100a67-kube-api-access-f8jvj\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.707406 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-extra-config-0\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.707427 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-2\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.707446 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-1\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.707507 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-0\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.711973 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51cf4d1e-7eb2-4882-bd28-c34e75100a67-kube-api-access-f8jvj" (OuterVolumeSpecName: "kube-api-access-f8jvj") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "kube-api-access-f8jvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.734370 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.734432 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.738859 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.739038 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.752352 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.752705 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.809628 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-0\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.809816 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-ssh-key-openstack-edpm-ipam\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.809954 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-combined-ca-bundle\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.810068 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-inventory\") pod \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\" (UID: \"51cf4d1e-7eb2-4882-bd28-c34e75100a67\") " Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.811672 4492 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.811698 4492 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.811710 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8jvj\" (UniqueName: \"kubernetes.io/projected/51cf4d1e-7eb2-4882-bd28-c34e75100a67-kube-api-access-f8jvj\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.811726 4492 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.811738 4492 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.811749 4492 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.811759 4492 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.813040 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.826804 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.828644 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.834290 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-inventory" (OuterVolumeSpecName: "inventory") pod "51cf4d1e-7eb2-4882-bd28-c34e75100a67" (UID: "51cf4d1e-7eb2-4882-bd28-c34e75100a67"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.913956 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.914064 4492 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.914128 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:13 crc kubenswrapper[4492]: I0220 07:17:13.914187 4492 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/51cf4d1e-7eb2-4882-bd28-c34e75100a67-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.311454 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" event={"ID":"51cf4d1e-7eb2-4882-bd28-c34e75100a67","Type":"ContainerDied","Data":"4192c14c9f5eaf7e52a42b5acd0dec75d1c5f5bf1f65b42d30b7d758d86cac87"} Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.311710 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4192c14c9f5eaf7e52a42b5acd0dec75d1c5f5bf1f65b42d30b7d758d86cac87" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.311534 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jqcw2" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.453661 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt"] Feb 20 07:17:14 crc kubenswrapper[4492]: E0220 07:17:14.454070 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="registry-server" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.454092 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="registry-server" Feb 20 07:17:14 crc kubenswrapper[4492]: E0220 07:17:14.454111 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51cf4d1e-7eb2-4882-bd28-c34e75100a67" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.454118 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="51cf4d1e-7eb2-4882-bd28-c34e75100a67" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 07:17:14 crc kubenswrapper[4492]: E0220 07:17:14.454125 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="extract-utilities" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.454131 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="extract-utilities" Feb 20 07:17:14 crc kubenswrapper[4492]: E0220 07:17:14.454145 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="extract-content" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.454150 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="extract-content" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.454343 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="df9394ea-402a-44ef-b1f6-365eef6d5f6a" containerName="registry-server" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.454359 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="51cf4d1e-7eb2-4882-bd28-c34e75100a67" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.455012 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.457961 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.458520 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.458788 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.458880 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.458971 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-l45gb" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.462382 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt"] Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.524535 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfcfm\" (UniqueName: \"kubernetes.io/projected/a79e5d8f-5072-4056-b1e3-b7853f563978-kube-api-access-hfcfm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.524772 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.524932 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.525054 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.525141 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.525249 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.525318 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.626335 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.626368 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.626529 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.626558 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.626588 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfcfm\" (UniqueName: \"kubernetes.io/projected/a79e5d8f-5072-4056-b1e3-b7853f563978-kube-api-access-hfcfm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.626608 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.627722 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.631956 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.631966 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.632861 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.633464 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.633627 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.635519 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.641850 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfcfm\" (UniqueName: \"kubernetes.io/projected/a79e5d8f-5072-4056-b1e3-b7853f563978-kube-api-access-hfcfm\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-npbqt\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:14 crc kubenswrapper[4492]: I0220 07:17:14.769873 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:17:15 crc kubenswrapper[4492]: I0220 07:17:15.235916 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt"] Feb 20 07:17:15 crc kubenswrapper[4492]: I0220 07:17:15.318205 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" event={"ID":"a79e5d8f-5072-4056-b1e3-b7853f563978","Type":"ContainerStarted","Data":"dd54470fa0c815429890165424b74debf706b8921692b947bddf16dcc344a198"} Feb 20 07:17:16 crc kubenswrapper[4492]: I0220 07:17:16.326159 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" event={"ID":"a79e5d8f-5072-4056-b1e3-b7853f563978","Type":"ContainerStarted","Data":"8de0d92d970ef910a0e81f78316d7fc7d8de8be999b467e370b5b8d8b179e2d0"} Feb 20 07:17:16 crc kubenswrapper[4492]: I0220 07:17:16.340321 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" podStartSLOduration=1.700227918 podStartE2EDuration="2.340303621s" podCreationTimestamp="2026-02-20 07:17:14 +0000 UTC" firstStartedPulling="2026-02-20 07:17:15.235978338 +0000 UTC m=+2192.007267316" lastFinishedPulling="2026-02-20 07:17:15.876054041 +0000 UTC m=+2192.647343019" observedRunningTime="2026-02-20 07:17:16.339417359 +0000 UTC m=+2193.110706337" watchObservedRunningTime="2026-02-20 07:17:16.340303621 +0000 UTC m=+2193.111592599" Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.311198 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.311836 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.311880 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.312354 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.312399 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" gracePeriod=600 Feb 20 07:17:39 crc kubenswrapper[4492]: E0220 07:17:39.430439 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.521609 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" exitCode=0 Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.521648 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97"} Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.521685 4492 scope.go:117] "RemoveContainer" containerID="766c6946637b4404c8c01854dc3221d37256795a37babb77f668a65574b03c74" Feb 20 07:17:39 crc kubenswrapper[4492]: I0220 07:17:39.522048 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:17:39 crc kubenswrapper[4492]: E0220 07:17:39.522276 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.557209 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:17:54 crc kubenswrapper[4492]: E0220 07:17:54.557951 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.674087 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ms7s8"] Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.682607 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.685985 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ms7s8"] Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.785454 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-utilities\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.785521 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-catalog-content\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.785538 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncd7n\" (UniqueName: \"kubernetes.io/projected/888acba4-cebe-4682-80dd-81af325f6707-kube-api-access-ncd7n\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.872870 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2p6kx"] Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.874652 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.882642 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2p6kx"] Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.887668 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-utilities\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.887727 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-catalog-content\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.887758 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncd7n\" (UniqueName: \"kubernetes.io/projected/888acba4-cebe-4682-80dd-81af325f6707-kube-api-access-ncd7n\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.887797 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxnsh\" (UniqueName: \"kubernetes.io/projected/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-kube-api-access-cxnsh\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.888012 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-catalog-content\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.888056 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-utilities\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.888081 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-utilities\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.888618 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-catalog-content\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.919339 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncd7n\" (UniqueName: \"kubernetes.io/projected/888acba4-cebe-4682-80dd-81af325f6707-kube-api-access-ncd7n\") pod \"redhat-operators-ms7s8\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.989325 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-catalog-content\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.989376 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-utilities\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.989460 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxnsh\" (UniqueName: \"kubernetes.io/projected/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-kube-api-access-cxnsh\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.989786 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-catalog-content\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:54 crc kubenswrapper[4492]: I0220 07:17:54.989836 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-utilities\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:55 crc kubenswrapper[4492]: I0220 07:17:55.005711 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:17:55 crc kubenswrapper[4492]: I0220 07:17:55.011602 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxnsh\" (UniqueName: \"kubernetes.io/projected/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-kube-api-access-cxnsh\") pod \"redhat-marketplace-2p6kx\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:55 crc kubenswrapper[4492]: I0220 07:17:55.194179 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:17:55 crc kubenswrapper[4492]: W0220 07:17:55.451965 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod888acba4_cebe_4682_80dd_81af325f6707.slice/crio-c9dd73ed732286eeba136edb3e035da3616e6c60ca6dc4ce2939c94e5d665a7e WatchSource:0}: Error finding container c9dd73ed732286eeba136edb3e035da3616e6c60ca6dc4ce2939c94e5d665a7e: Status 404 returned error can't find the container with id c9dd73ed732286eeba136edb3e035da3616e6c60ca6dc4ce2939c94e5d665a7e Feb 20 07:17:55 crc kubenswrapper[4492]: I0220 07:17:55.459897 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ms7s8"] Feb 20 07:17:55 crc kubenswrapper[4492]: I0220 07:17:55.627951 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2p6kx"] Feb 20 07:17:55 crc kubenswrapper[4492]: I0220 07:17:55.638645 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ms7s8" event={"ID":"888acba4-cebe-4682-80dd-81af325f6707","Type":"ContainerStarted","Data":"c9dd73ed732286eeba136edb3e035da3616e6c60ca6dc4ce2939c94e5d665a7e"} Feb 20 07:17:56 crc kubenswrapper[4492]: I0220 07:17:56.648043 4492 generic.go:334] "Generic (PLEG): container finished" podID="888acba4-cebe-4682-80dd-81af325f6707" containerID="07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60" exitCode=0 Feb 20 07:17:56 crc kubenswrapper[4492]: I0220 07:17:56.648105 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ms7s8" event={"ID":"888acba4-cebe-4682-80dd-81af325f6707","Type":"ContainerDied","Data":"07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60"} Feb 20 07:17:56 crc kubenswrapper[4492]: I0220 07:17:56.649877 4492 generic.go:334] "Generic (PLEG): container finished" podID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerID="2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f" exitCode=0 Feb 20 07:17:56 crc kubenswrapper[4492]: I0220 07:17:56.649913 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2p6kx" event={"ID":"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc","Type":"ContainerDied","Data":"2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f"} Feb 20 07:17:56 crc kubenswrapper[4492]: I0220 07:17:56.649935 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2p6kx" event={"ID":"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc","Type":"ContainerStarted","Data":"03d0058237ff845496ef367037cb6f36c0219fa6850c01d0b826c92e1b204bac"} Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.267754 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c22x4"] Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.271020 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.277879 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c22x4"] Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.441413 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-utilities\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.444116 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-catalog-content\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.444182 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnn9b\" (UniqueName: \"kubernetes.io/projected/8e973200-db57-4851-8c39-c854dc791b76-kube-api-access-vnn9b\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.546450 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-catalog-content\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.546508 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnn9b\" (UniqueName: \"kubernetes.io/projected/8e973200-db57-4851-8c39-c854dc791b76-kube-api-access-vnn9b\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.546636 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-utilities\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.547013 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-catalog-content\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.547076 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-utilities\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.583823 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnn9b\" (UniqueName: \"kubernetes.io/projected/8e973200-db57-4851-8c39-c854dc791b76-kube-api-access-vnn9b\") pod \"community-operators-c22x4\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.587265 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.667638 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2p6kx" event={"ID":"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc","Type":"ContainerStarted","Data":"a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd"} Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.674648 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ms7s8" event={"ID":"888acba4-cebe-4682-80dd-81af325f6707","Type":"ContainerStarted","Data":"25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff"} Feb 20 07:17:57 crc kubenswrapper[4492]: I0220 07:17:57.893830 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c22x4"] Feb 20 07:17:58 crc kubenswrapper[4492]: I0220 07:17:58.683935 4492 generic.go:334] "Generic (PLEG): container finished" podID="8e973200-db57-4851-8c39-c854dc791b76" containerID="0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253" exitCode=0 Feb 20 07:17:58 crc kubenswrapper[4492]: I0220 07:17:58.684334 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c22x4" event={"ID":"8e973200-db57-4851-8c39-c854dc791b76","Type":"ContainerDied","Data":"0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253"} Feb 20 07:17:58 crc kubenswrapper[4492]: I0220 07:17:58.684380 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c22x4" event={"ID":"8e973200-db57-4851-8c39-c854dc791b76","Type":"ContainerStarted","Data":"a55f55278e8879efcf87c1b655a38b8502708b1c59f99dd9aebff80d43634860"} Feb 20 07:17:59 crc kubenswrapper[4492]: I0220 07:17:59.711312 4492 generic.go:334] "Generic (PLEG): container finished" podID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerID="a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd" exitCode=0 Feb 20 07:17:59 crc kubenswrapper[4492]: I0220 07:17:59.711428 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2p6kx" event={"ID":"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc","Type":"ContainerDied","Data":"a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd"} Feb 20 07:17:59 crc kubenswrapper[4492]: I0220 07:17:59.716727 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c22x4" event={"ID":"8e973200-db57-4851-8c39-c854dc791b76","Type":"ContainerStarted","Data":"316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f"} Feb 20 07:18:00 crc kubenswrapper[4492]: I0220 07:18:00.724346 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2p6kx" event={"ID":"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc","Type":"ContainerStarted","Data":"8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d"} Feb 20 07:18:00 crc kubenswrapper[4492]: I0220 07:18:00.726825 4492 generic.go:334] "Generic (PLEG): container finished" podID="888acba4-cebe-4682-80dd-81af325f6707" containerID="25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff" exitCode=0 Feb 20 07:18:00 crc kubenswrapper[4492]: I0220 07:18:00.726899 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ms7s8" event={"ID":"888acba4-cebe-4682-80dd-81af325f6707","Type":"ContainerDied","Data":"25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff"} Feb 20 07:18:00 crc kubenswrapper[4492]: I0220 07:18:00.745107 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2p6kx" podStartSLOduration=3.180332423 podStartE2EDuration="6.745094757s" podCreationTimestamp="2026-02-20 07:17:54 +0000 UTC" firstStartedPulling="2026-02-20 07:17:56.651174989 +0000 UTC m=+2233.422463967" lastFinishedPulling="2026-02-20 07:18:00.215937323 +0000 UTC m=+2236.987226301" observedRunningTime="2026-02-20 07:18:00.740349158 +0000 UTC m=+2237.511638136" watchObservedRunningTime="2026-02-20 07:18:00.745094757 +0000 UTC m=+2237.516383734" Feb 20 07:18:01 crc kubenswrapper[4492]: I0220 07:18:01.739019 4492 generic.go:334] "Generic (PLEG): container finished" podID="8e973200-db57-4851-8c39-c854dc791b76" containerID="316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f" exitCode=0 Feb 20 07:18:01 crc kubenswrapper[4492]: I0220 07:18:01.739126 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c22x4" event={"ID":"8e973200-db57-4851-8c39-c854dc791b76","Type":"ContainerDied","Data":"316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f"} Feb 20 07:18:01 crc kubenswrapper[4492]: I0220 07:18:01.741652 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ms7s8" event={"ID":"888acba4-cebe-4682-80dd-81af325f6707","Type":"ContainerStarted","Data":"a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9"} Feb 20 07:18:01 crc kubenswrapper[4492]: I0220 07:18:01.781799 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ms7s8" podStartSLOduration=3.245429238 podStartE2EDuration="7.781780338s" podCreationTimestamp="2026-02-20 07:17:54 +0000 UTC" firstStartedPulling="2026-02-20 07:17:56.649763147 +0000 UTC m=+2233.421052125" lastFinishedPulling="2026-02-20 07:18:01.186114247 +0000 UTC m=+2237.957403225" observedRunningTime="2026-02-20 07:18:01.771270702 +0000 UTC m=+2238.542559679" watchObservedRunningTime="2026-02-20 07:18:01.781780338 +0000 UTC m=+2238.553069316" Feb 20 07:18:02 crc kubenswrapper[4492]: I0220 07:18:02.750201 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c22x4" event={"ID":"8e973200-db57-4851-8c39-c854dc791b76","Type":"ContainerStarted","Data":"3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718"} Feb 20 07:18:05 crc kubenswrapper[4492]: I0220 07:18:05.006939 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:18:05 crc kubenswrapper[4492]: I0220 07:18:05.007649 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:18:05 crc kubenswrapper[4492]: I0220 07:18:05.192454 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:18:05 crc kubenswrapper[4492]: I0220 07:18:05.192794 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:18:05 crc kubenswrapper[4492]: I0220 07:18:05.237347 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:18:05 crc kubenswrapper[4492]: I0220 07:18:05.256908 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c22x4" podStartSLOduration=4.7720343100000004 podStartE2EDuration="8.256892568s" podCreationTimestamp="2026-02-20 07:17:57 +0000 UTC" firstStartedPulling="2026-02-20 07:17:58.68704972 +0000 UTC m=+2235.458338688" lastFinishedPulling="2026-02-20 07:18:02.171907968 +0000 UTC m=+2238.943196946" observedRunningTime="2026-02-20 07:18:02.775321778 +0000 UTC m=+2239.546610757" watchObservedRunningTime="2026-02-20 07:18:05.256892568 +0000 UTC m=+2242.028181547" Feb 20 07:18:05 crc kubenswrapper[4492]: I0220 07:18:05.843793 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:18:06 crc kubenswrapper[4492]: I0220 07:18:06.046373 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ms7s8" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="registry-server" probeResult="failure" output=< Feb 20 07:18:06 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:18:06 crc kubenswrapper[4492]: > Feb 20 07:18:06 crc kubenswrapper[4492]: I0220 07:18:06.556887 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:18:06 crc kubenswrapper[4492]: E0220 07:18:06.557396 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:18:06 crc kubenswrapper[4492]: I0220 07:18:06.666530 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2p6kx"] Feb 20 07:18:07 crc kubenswrapper[4492]: I0220 07:18:07.588454 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:18:07 crc kubenswrapper[4492]: I0220 07:18:07.588633 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:18:07 crc kubenswrapper[4492]: I0220 07:18:07.626074 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:18:07 crc kubenswrapper[4492]: I0220 07:18:07.802863 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2p6kx" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="registry-server" containerID="cri-o://8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d" gracePeriod=2 Feb 20 07:18:07 crc kubenswrapper[4492]: I0220 07:18:07.841428 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.231975 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.271754 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-catalog-content\") pod \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.271799 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-utilities\") pod \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.272150 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxnsh\" (UniqueName: \"kubernetes.io/projected/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-kube-api-access-cxnsh\") pod \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\" (UID: \"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc\") " Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.272583 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-utilities" (OuterVolumeSpecName: "utilities") pod "2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" (UID: "2dbbef13-84bc-43e1-99ca-7e0bfc7386cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.273432 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.279527 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-kube-api-access-cxnsh" (OuterVolumeSpecName: "kube-api-access-cxnsh") pod "2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" (UID: "2dbbef13-84bc-43e1-99ca-7e0bfc7386cc"). InnerVolumeSpecName "kube-api-access-cxnsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.293056 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" (UID: "2dbbef13-84bc-43e1-99ca-7e0bfc7386cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.375554 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxnsh\" (UniqueName: \"kubernetes.io/projected/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-kube-api-access-cxnsh\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.375594 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.813272 4492 generic.go:334] "Generic (PLEG): container finished" podID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerID="8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d" exitCode=0 Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.813379 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2p6kx" event={"ID":"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc","Type":"ContainerDied","Data":"8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d"} Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.813461 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2p6kx" event={"ID":"2dbbef13-84bc-43e1-99ca-7e0bfc7386cc","Type":"ContainerDied","Data":"03d0058237ff845496ef367037cb6f36c0219fa6850c01d0b826c92e1b204bac"} Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.813520 4492 scope.go:117] "RemoveContainer" containerID="8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.813394 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2p6kx" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.830941 4492 scope.go:117] "RemoveContainer" containerID="a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.847248 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2p6kx"] Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.855921 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2p6kx"] Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.858754 4492 scope.go:117] "RemoveContainer" containerID="2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.901092 4492 scope.go:117] "RemoveContainer" containerID="8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d" Feb 20 07:18:08 crc kubenswrapper[4492]: E0220 07:18:08.901750 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d\": container with ID starting with 8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d not found: ID does not exist" containerID="8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.901808 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d"} err="failed to get container status \"8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d\": rpc error: code = NotFound desc = could not find container \"8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d\": container with ID starting with 8a8bfc79caa5babdcebda47e3fc840779e64d14df07390265f846a254befbf1d not found: ID does not exist" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.901850 4492 scope.go:117] "RemoveContainer" containerID="a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd" Feb 20 07:18:08 crc kubenswrapper[4492]: E0220 07:18:08.902567 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd\": container with ID starting with a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd not found: ID does not exist" containerID="a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.902626 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd"} err="failed to get container status \"a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd\": rpc error: code = NotFound desc = could not find container \"a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd\": container with ID starting with a65954910460285f51e251a992fe0c6cfaf5d12178732f303f984b23bbea50cd not found: ID does not exist" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.902660 4492 scope.go:117] "RemoveContainer" containerID="2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f" Feb 20 07:18:08 crc kubenswrapper[4492]: E0220 07:18:08.903138 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f\": container with ID starting with 2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f not found: ID does not exist" containerID="2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f" Feb 20 07:18:08 crc kubenswrapper[4492]: I0220 07:18:08.903179 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f"} err="failed to get container status \"2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f\": rpc error: code = NotFound desc = could not find container \"2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f\": container with ID starting with 2fa47f5c4c48096abc91507fdfa7d12883f100e5b82899a02bd0a5f4e8de0b5f not found: ID does not exist" Feb 20 07:18:09 crc kubenswrapper[4492]: I0220 07:18:09.565743 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" path="/var/lib/kubelet/pods/2dbbef13-84bc-43e1-99ca-7e0bfc7386cc/volumes" Feb 20 07:18:09 crc kubenswrapper[4492]: I0220 07:18:09.864846 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c22x4"] Feb 20 07:18:09 crc kubenswrapper[4492]: I0220 07:18:09.865587 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c22x4" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="registry-server" containerID="cri-o://3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718" gracePeriod=2 Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.297559 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.421047 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-catalog-content\") pod \"8e973200-db57-4851-8c39-c854dc791b76\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.421168 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnn9b\" (UniqueName: \"kubernetes.io/projected/8e973200-db57-4851-8c39-c854dc791b76-kube-api-access-vnn9b\") pod \"8e973200-db57-4851-8c39-c854dc791b76\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.421197 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-utilities\") pod \"8e973200-db57-4851-8c39-c854dc791b76\" (UID: \"8e973200-db57-4851-8c39-c854dc791b76\") " Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.422281 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-utilities" (OuterVolumeSpecName: "utilities") pod "8e973200-db57-4851-8c39-c854dc791b76" (UID: "8e973200-db57-4851-8c39-c854dc791b76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.435745 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e973200-db57-4851-8c39-c854dc791b76-kube-api-access-vnn9b" (OuterVolumeSpecName: "kube-api-access-vnn9b") pod "8e973200-db57-4851-8c39-c854dc791b76" (UID: "8e973200-db57-4851-8c39-c854dc791b76"). InnerVolumeSpecName "kube-api-access-vnn9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.470384 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e973200-db57-4851-8c39-c854dc791b76" (UID: "8e973200-db57-4851-8c39-c854dc791b76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.523661 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.524008 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e973200-db57-4851-8c39-c854dc791b76-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.524094 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnn9b\" (UniqueName: \"kubernetes.io/projected/8e973200-db57-4851-8c39-c854dc791b76-kube-api-access-vnn9b\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.840670 4492 generic.go:334] "Generic (PLEG): container finished" podID="8e973200-db57-4851-8c39-c854dc791b76" containerID="3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718" exitCode=0 Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.840754 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c22x4" event={"ID":"8e973200-db57-4851-8c39-c854dc791b76","Type":"ContainerDied","Data":"3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718"} Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.840775 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c22x4" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.840820 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c22x4" event={"ID":"8e973200-db57-4851-8c39-c854dc791b76","Type":"ContainerDied","Data":"a55f55278e8879efcf87c1b655a38b8502708b1c59f99dd9aebff80d43634860"} Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.840857 4492 scope.go:117] "RemoveContainer" containerID="3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.868854 4492 scope.go:117] "RemoveContainer" containerID="316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.871777 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c22x4"] Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.880807 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c22x4"] Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.893806 4492 scope.go:117] "RemoveContainer" containerID="0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.928061 4492 scope.go:117] "RemoveContainer" containerID="3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718" Feb 20 07:18:10 crc kubenswrapper[4492]: E0220 07:18:10.930139 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718\": container with ID starting with 3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718 not found: ID does not exist" containerID="3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.930167 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718"} err="failed to get container status \"3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718\": rpc error: code = NotFound desc = could not find container \"3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718\": container with ID starting with 3a2fb3bcad3734e3f6b1b8779638a058709d631a1bba0d2ad4a369bea8b41718 not found: ID does not exist" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.930188 4492 scope.go:117] "RemoveContainer" containerID="316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f" Feb 20 07:18:10 crc kubenswrapper[4492]: E0220 07:18:10.930515 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f\": container with ID starting with 316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f not found: ID does not exist" containerID="316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.930541 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f"} err="failed to get container status \"316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f\": rpc error: code = NotFound desc = could not find container \"316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f\": container with ID starting with 316b6d32613db39f9cc4fb4224e6a2994879c46572d07d94cddd8f9d4b81078f not found: ID does not exist" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.930557 4492 scope.go:117] "RemoveContainer" containerID="0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253" Feb 20 07:18:10 crc kubenswrapper[4492]: E0220 07:18:10.931238 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253\": container with ID starting with 0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253 not found: ID does not exist" containerID="0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253" Feb 20 07:18:10 crc kubenswrapper[4492]: I0220 07:18:10.931258 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253"} err="failed to get container status \"0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253\": rpc error: code = NotFound desc = could not find container \"0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253\": container with ID starting with 0743da27af2cce1fedc0c3cc471de50ec6de400149b6756a5b0d1e6bf49a0253 not found: ID does not exist" Feb 20 07:18:11 crc kubenswrapper[4492]: I0220 07:18:11.565365 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e973200-db57-4851-8c39-c854dc791b76" path="/var/lib/kubelet/pods/8e973200-db57-4851-8c39-c854dc791b76/volumes" Feb 20 07:18:16 crc kubenswrapper[4492]: I0220 07:18:16.043175 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ms7s8" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="registry-server" probeResult="failure" output=< Feb 20 07:18:16 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:18:16 crc kubenswrapper[4492]: > Feb 20 07:18:17 crc kubenswrapper[4492]: I0220 07:18:17.557703 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:18:17 crc kubenswrapper[4492]: E0220 07:18:17.558269 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:18:25 crc kubenswrapper[4492]: I0220 07:18:25.042409 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:18:25 crc kubenswrapper[4492]: I0220 07:18:25.088698 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:18:25 crc kubenswrapper[4492]: I0220 07:18:25.884033 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ms7s8"] Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.005152 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ms7s8" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="registry-server" containerID="cri-o://a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9" gracePeriod=2 Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.433614 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.610905 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-utilities\") pod \"888acba4-cebe-4682-80dd-81af325f6707\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.611540 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-utilities" (OuterVolumeSpecName: "utilities") pod "888acba4-cebe-4682-80dd-81af325f6707" (UID: "888acba4-cebe-4682-80dd-81af325f6707"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.611929 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncd7n\" (UniqueName: \"kubernetes.io/projected/888acba4-cebe-4682-80dd-81af325f6707-kube-api-access-ncd7n\") pod \"888acba4-cebe-4682-80dd-81af325f6707\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.612002 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-catalog-content\") pod \"888acba4-cebe-4682-80dd-81af325f6707\" (UID: \"888acba4-cebe-4682-80dd-81af325f6707\") " Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.613869 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.618653 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/888acba4-cebe-4682-80dd-81af325f6707-kube-api-access-ncd7n" (OuterVolumeSpecName: "kube-api-access-ncd7n") pod "888acba4-cebe-4682-80dd-81af325f6707" (UID: "888acba4-cebe-4682-80dd-81af325f6707"). InnerVolumeSpecName "kube-api-access-ncd7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.720156 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "888acba4-cebe-4682-80dd-81af325f6707" (UID: "888acba4-cebe-4682-80dd-81af325f6707"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.720564 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncd7n\" (UniqueName: \"kubernetes.io/projected/888acba4-cebe-4682-80dd-81af325f6707-kube-api-access-ncd7n\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:27 crc kubenswrapper[4492]: I0220 07:18:27.720695 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/888acba4-cebe-4682-80dd-81af325f6707-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.023313 4492 generic.go:334] "Generic (PLEG): container finished" podID="888acba4-cebe-4682-80dd-81af325f6707" containerID="a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9" exitCode=0 Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.023408 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ms7s8" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.023398 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ms7s8" event={"ID":"888acba4-cebe-4682-80dd-81af325f6707","Type":"ContainerDied","Data":"a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9"} Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.023531 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ms7s8" event={"ID":"888acba4-cebe-4682-80dd-81af325f6707","Type":"ContainerDied","Data":"c9dd73ed732286eeba136edb3e035da3616e6c60ca6dc4ce2939c94e5d665a7e"} Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.023566 4492 scope.go:117] "RemoveContainer" containerID="a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.086497 4492 scope.go:117] "RemoveContainer" containerID="25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.108628 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ms7s8"] Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.114986 4492 scope.go:117] "RemoveContainer" containerID="07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.119027 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ms7s8"] Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.150982 4492 scope.go:117] "RemoveContainer" containerID="a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9" Feb 20 07:18:28 crc kubenswrapper[4492]: E0220 07:18:28.151350 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9\": container with ID starting with a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9 not found: ID does not exist" containerID="a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.151405 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9"} err="failed to get container status \"a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9\": rpc error: code = NotFound desc = could not find container \"a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9\": container with ID starting with a9b002858418b048931a3ee75d12c9409aae573c6cb548fdf44ec87d440db9e9 not found: ID does not exist" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.151436 4492 scope.go:117] "RemoveContainer" containerID="25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff" Feb 20 07:18:28 crc kubenswrapper[4492]: E0220 07:18:28.152026 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff\": container with ID starting with 25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff not found: ID does not exist" containerID="25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.152061 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff"} err="failed to get container status \"25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff\": rpc error: code = NotFound desc = could not find container \"25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff\": container with ID starting with 25f345ee9c5fbffc67c07da5292b193c6cf59e22548ef18740b0bc49456b76ff not found: ID does not exist" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.152087 4492 scope.go:117] "RemoveContainer" containerID="07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60" Feb 20 07:18:28 crc kubenswrapper[4492]: E0220 07:18:28.152521 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60\": container with ID starting with 07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60 not found: ID does not exist" containerID="07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60" Feb 20 07:18:28 crc kubenswrapper[4492]: I0220 07:18:28.152550 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60"} err="failed to get container status \"07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60\": rpc error: code = NotFound desc = could not find container \"07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60\": container with ID starting with 07cb5ef7a98adb8d9092b591b3fdcc13f36ddf6a2f8ed84c8f59a8d28f89be60 not found: ID does not exist" Feb 20 07:18:29 crc kubenswrapper[4492]: I0220 07:18:29.567434 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="888acba4-cebe-4682-80dd-81af325f6707" path="/var/lib/kubelet/pods/888acba4-cebe-4682-80dd-81af325f6707/volumes" Feb 20 07:18:32 crc kubenswrapper[4492]: I0220 07:18:32.557424 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:18:32 crc kubenswrapper[4492]: E0220 07:18:32.558100 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:18:47 crc kubenswrapper[4492]: I0220 07:18:47.557072 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:18:47 crc kubenswrapper[4492]: E0220 07:18:47.558038 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:18:58 crc kubenswrapper[4492]: I0220 07:18:58.556874 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:18:58 crc kubenswrapper[4492]: E0220 07:18:58.557626 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:19:11 crc kubenswrapper[4492]: I0220 07:19:11.557247 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:19:11 crc kubenswrapper[4492]: E0220 07:19:11.558246 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:19:22 crc kubenswrapper[4492]: I0220 07:19:22.557629 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:19:22 crc kubenswrapper[4492]: E0220 07:19:22.558683 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:19:29 crc kubenswrapper[4492]: I0220 07:19:29.536342 4492 generic.go:334] "Generic (PLEG): container finished" podID="a79e5d8f-5072-4056-b1e3-b7853f563978" containerID="8de0d92d970ef910a0e81f78316d7fc7d8de8be999b467e370b5b8d8b179e2d0" exitCode=0 Feb 20 07:19:29 crc kubenswrapper[4492]: I0220 07:19:29.536436 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" event={"ID":"a79e5d8f-5072-4056-b1e3-b7853f563978","Type":"ContainerDied","Data":"8de0d92d970ef910a0e81f78316d7fc7d8de8be999b467e370b5b8d8b179e2d0"} Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.894661 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.991160 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfcfm\" (UniqueName: \"kubernetes.io/projected/a79e5d8f-5072-4056-b1e3-b7853f563978-kube-api-access-hfcfm\") pod \"a79e5d8f-5072-4056-b1e3-b7853f563978\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.991237 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-inventory\") pod \"a79e5d8f-5072-4056-b1e3-b7853f563978\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.991269 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-0\") pod \"a79e5d8f-5072-4056-b1e3-b7853f563978\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.991321 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-2\") pod \"a79e5d8f-5072-4056-b1e3-b7853f563978\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.991446 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-telemetry-combined-ca-bundle\") pod \"a79e5d8f-5072-4056-b1e3-b7853f563978\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.991505 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-1\") pod \"a79e5d8f-5072-4056-b1e3-b7853f563978\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.991566 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ssh-key-openstack-edpm-ipam\") pod \"a79e5d8f-5072-4056-b1e3-b7853f563978\" (UID: \"a79e5d8f-5072-4056-b1e3-b7853f563978\") " Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.998774 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a79e5d8f-5072-4056-b1e3-b7853f563978-kube-api-access-hfcfm" (OuterVolumeSpecName: "kube-api-access-hfcfm") pod "a79e5d8f-5072-4056-b1e3-b7853f563978" (UID: "a79e5d8f-5072-4056-b1e3-b7853f563978"). InnerVolumeSpecName "kube-api-access-hfcfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:19:30 crc kubenswrapper[4492]: I0220 07:19:30.999456 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a79e5d8f-5072-4056-b1e3-b7853f563978" (UID: "a79e5d8f-5072-4056-b1e3-b7853f563978"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.018825 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-inventory" (OuterVolumeSpecName: "inventory") pod "a79e5d8f-5072-4056-b1e3-b7853f563978" (UID: "a79e5d8f-5072-4056-b1e3-b7853f563978"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.019328 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "a79e5d8f-5072-4056-b1e3-b7853f563978" (UID: "a79e5d8f-5072-4056-b1e3-b7853f563978"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.019899 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "a79e5d8f-5072-4056-b1e3-b7853f563978" (UID: "a79e5d8f-5072-4056-b1e3-b7853f563978"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.021457 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a79e5d8f-5072-4056-b1e3-b7853f563978" (UID: "a79e5d8f-5072-4056-b1e3-b7853f563978"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.025902 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "a79e5d8f-5072-4056-b1e3-b7853f563978" (UID: "a79e5d8f-5072-4056-b1e3-b7853f563978"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.094111 4492 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.094137 4492 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.094148 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.094157 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfcfm\" (UniqueName: \"kubernetes.io/projected/a79e5d8f-5072-4056-b1e3-b7853f563978-kube-api-access-hfcfm\") on node \"crc\" DevicePath \"\"" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.094172 4492 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-inventory\") on node \"crc\" DevicePath \"\"" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.094182 4492 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.094191 4492 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/a79e5d8f-5072-4056-b1e3-b7853f563978-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.555987 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.556014 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-npbqt" event={"ID":"a79e5d8f-5072-4056-b1e3-b7853f563978","Type":"ContainerDied","Data":"dd54470fa0c815429890165424b74debf706b8921692b947bddf16dcc344a198"} Feb 20 07:19:31 crc kubenswrapper[4492]: I0220 07:19:31.556300 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd54470fa0c815429890165424b74debf706b8921692b947bddf16dcc344a198" Feb 20 07:19:33 crc kubenswrapper[4492]: I0220 07:19:33.562085 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:19:33 crc kubenswrapper[4492]: E0220 07:19:33.562753 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:19:47 crc kubenswrapper[4492]: I0220 07:19:47.557411 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:19:47 crc kubenswrapper[4492]: E0220 07:19:47.558345 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:20:01 crc kubenswrapper[4492]: I0220 07:20:01.557709 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:20:01 crc kubenswrapper[4492]: E0220 07:20:01.558730 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:20:16 crc kubenswrapper[4492]: I0220 07:20:16.556378 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:20:16 crc kubenswrapper[4492]: E0220 07:20:16.557256 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.432403 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.433960 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.433985 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434017 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="extract-content" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434025 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="extract-content" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434045 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434051 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434062 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="extract-utilities" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434069 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="extract-utilities" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434082 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="extract-content" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434088 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="extract-content" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434106 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="extract-content" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434111 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="extract-content" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434120 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="extract-utilities" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434127 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="extract-utilities" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434144 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79e5d8f-5072-4056-b1e3-b7853f563978" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434152 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79e5d8f-5072-4056-b1e3-b7853f563978" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434161 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="extract-utilities" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434167 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="extract-utilities" Feb 20 07:20:26 crc kubenswrapper[4492]: E0220 07:20:26.434183 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434189 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434564 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e973200-db57-4851-8c39-c854dc791b76" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434588 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a79e5d8f-5072-4056-b1e3-b7853f563978" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434614 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="888acba4-cebe-4682-80dd-81af325f6707" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.434627 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dbbef13-84bc-43e1-99ca-7e0bfc7386cc" containerName="registry-server" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.435818 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.437812 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.437929 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.438301 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.438409 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pdp6n" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.448712 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.524977 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.525095 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.525258 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626592 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626639 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626663 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626729 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626774 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626806 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626832 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626858 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.626891 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5ff\" (UniqueName: \"kubernetes.io/projected/419ec593-81d8-44f8-9d19-5df9ebf183e1-kube-api-access-xm5ff\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.629045 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.629235 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.632428 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.728661 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.729009 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.729106 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.729150 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.729212 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5ff\" (UniqueName: \"kubernetes.io/projected/419ec593-81d8-44f8-9d19-5df9ebf183e1-kube-api-access-xm5ff\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.729327 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.729357 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.730466 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.730897 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.732932 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.736346 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.749993 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5ff\" (UniqueName: \"kubernetes.io/projected/419ec593-81d8-44f8-9d19-5df9ebf183e1-kube-api-access-xm5ff\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:26 crc kubenswrapper[4492]: I0220 07:20:26.758795 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:27 crc kubenswrapper[4492]: I0220 07:20:27.056884 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 07:20:27 crc kubenswrapper[4492]: I0220 07:20:27.600454 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Feb 20 07:20:27 crc kubenswrapper[4492]: I0220 07:20:27.610587 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:20:28 crc kubenswrapper[4492]: I0220 07:20:28.059107 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"419ec593-81d8-44f8-9d19-5df9ebf183e1","Type":"ContainerStarted","Data":"a94ba04d5aabd2da6882f3d8a9211007bb2cc78aeb1dfab60e62ba84b39e265a"} Feb 20 07:20:31 crc kubenswrapper[4492]: I0220 07:20:31.558017 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:20:31 crc kubenswrapper[4492]: E0220 07:20:31.559077 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:20:45 crc kubenswrapper[4492]: I0220 07:20:45.557904 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:20:45 crc kubenswrapper[4492]: E0220 07:20:45.558523 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:21:00 crc kubenswrapper[4492]: I0220 07:21:00.557469 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:21:00 crc kubenswrapper[4492]: E0220 07:21:00.560048 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:21:13 crc kubenswrapper[4492]: I0220 07:21:13.563070 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:21:13 crc kubenswrapper[4492]: E0220 07:21:13.563843 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:21:13 crc kubenswrapper[4492]: E0220 07:21:13.601187 4492 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:f0473f3e18dd17d7021c02e991298923" Feb 20 07:21:13 crc kubenswrapper[4492]: E0220 07:21:13.601246 4492 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:f0473f3e18dd17d7021c02e991298923" Feb 20 07:21:13 crc kubenswrapper[4492]: E0220 07:21:13.603242 4492 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:f0473f3e18dd17d7021c02e991298923,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xm5ff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-multi-thread-testing_openstack(419ec593-81d8-44f8-9d19-5df9ebf183e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 20 07:21:13 crc kubenswrapper[4492]: E0220 07:21:13.604533 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podUID="419ec593-81d8-44f8-9d19-5df9ebf183e1" Feb 20 07:21:14 crc kubenswrapper[4492]: E0220 07:21:14.542260 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:f0473f3e18dd17d7021c02e991298923\\\"\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podUID="419ec593-81d8-44f8-9d19-5df9ebf183e1" Feb 20 07:21:25 crc kubenswrapper[4492]: I0220 07:21:25.556665 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:21:25 crc kubenswrapper[4492]: E0220 07:21:25.557552 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:21:27 crc kubenswrapper[4492]: I0220 07:21:27.214095 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 20 07:21:28 crc kubenswrapper[4492]: I0220 07:21:28.683228 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"419ec593-81d8-44f8-9d19-5df9ebf183e1","Type":"ContainerStarted","Data":"01d4922c7c8ea5fe9cf1952e2a8cbb04bac6d0f58f7b3cf534717314f9c3f53f"} Feb 20 07:21:28 crc kubenswrapper[4492]: I0220 07:21:28.711740 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podStartSLOduration=4.112872576 podStartE2EDuration="1m3.711716866s" podCreationTimestamp="2026-02-20 07:20:25 +0000 UTC" firstStartedPulling="2026-02-20 07:20:27.610305544 +0000 UTC m=+2384.381594521" lastFinishedPulling="2026-02-20 07:21:27.209149833 +0000 UTC m=+2443.980438811" observedRunningTime="2026-02-20 07:21:28.702870004 +0000 UTC m=+2445.474158982" watchObservedRunningTime="2026-02-20 07:21:28.711716866 +0000 UTC m=+2445.483005845" Feb 20 07:21:37 crc kubenswrapper[4492]: I0220 07:21:37.557563 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:21:37 crc kubenswrapper[4492]: E0220 07:21:37.558557 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:21:52 crc kubenswrapper[4492]: I0220 07:21:52.557137 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:21:52 crc kubenswrapper[4492]: E0220 07:21:52.558672 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:22:04 crc kubenswrapper[4492]: I0220 07:22:04.556915 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:22:04 crc kubenswrapper[4492]: E0220 07:22:04.557777 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:22:17 crc kubenswrapper[4492]: I0220 07:22:17.560029 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:22:17 crc kubenswrapper[4492]: E0220 07:22:17.561296 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:22:32 crc kubenswrapper[4492]: I0220 07:22:32.558430 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:22:32 crc kubenswrapper[4492]: E0220 07:22:32.560412 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:22:43 crc kubenswrapper[4492]: I0220 07:22:43.562444 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:22:44 crc kubenswrapper[4492]: I0220 07:22:44.349730 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"ea2dd75e96a336a1f60c20e23825e1bcd69e3e675203868cf884eb3d626bc2c4"} Feb 20 07:25:09 crc kubenswrapper[4492]: I0220 07:25:09.313036 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:25:09 crc kubenswrapper[4492]: I0220 07:25:09.316548 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:25:39 crc kubenswrapper[4492]: I0220 07:25:39.311421 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:25:39 crc kubenswrapper[4492]: I0220 07:25:39.312142 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:26:09 crc kubenswrapper[4492]: I0220 07:26:09.311209 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:26:09 crc kubenswrapper[4492]: I0220 07:26:09.311774 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:26:09 crc kubenswrapper[4492]: I0220 07:26:09.312851 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:26:09 crc kubenswrapper[4492]: I0220 07:26:09.314382 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ea2dd75e96a336a1f60c20e23825e1bcd69e3e675203868cf884eb3d626bc2c4"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:26:09 crc kubenswrapper[4492]: I0220 07:26:09.314860 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://ea2dd75e96a336a1f60c20e23825e1bcd69e3e675203868cf884eb3d626bc2c4" gracePeriod=600 Feb 20 07:26:10 crc kubenswrapper[4492]: I0220 07:26:10.139764 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="ea2dd75e96a336a1f60c20e23825e1bcd69e3e675203868cf884eb3d626bc2c4" exitCode=0 Feb 20 07:26:10 crc kubenswrapper[4492]: I0220 07:26:10.139934 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"ea2dd75e96a336a1f60c20e23825e1bcd69e3e675203868cf884eb3d626bc2c4"} Feb 20 07:26:10 crc kubenswrapper[4492]: I0220 07:26:10.140516 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9"} Feb 20 07:26:10 crc kubenswrapper[4492]: I0220 07:26:10.142178 4492 scope.go:117] "RemoveContainer" containerID="06734348a4f8d5a35fdcce0a5e34112a166f3f22b98c1026e06dd3bf880bfe97" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.777193 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hnw7z"] Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.786783 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.858325 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hnw7z"] Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.858998 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-utilities\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.859060 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-catalog-content\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.859209 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xzw\" (UniqueName: \"kubernetes.io/projected/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-kube-api-access-b7xzw\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.961589 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-catalog-content\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.961814 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xzw\" (UniqueName: \"kubernetes.io/projected/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-kube-api-access-b7xzw\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.962036 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-utilities\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.964310 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-catalog-content\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.965351 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-utilities\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:11 crc kubenswrapper[4492]: I0220 07:27:11.986223 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xzw\" (UniqueName: \"kubernetes.io/projected/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-kube-api-access-b7xzw\") pod \"certified-operators-hnw7z\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:12 crc kubenswrapper[4492]: I0220 07:27:12.108867 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:12 crc kubenswrapper[4492]: I0220 07:27:12.892377 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hnw7z"] Feb 20 07:27:13 crc kubenswrapper[4492]: I0220 07:27:13.641984 4492 generic.go:334] "Generic (PLEG): container finished" podID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerID="7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6" exitCode=0 Feb 20 07:27:13 crc kubenswrapper[4492]: I0220 07:27:13.642093 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnw7z" event={"ID":"0311469a-95b9-4079-8a9f-e4d4a67e3b5f","Type":"ContainerDied","Data":"7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6"} Feb 20 07:27:13 crc kubenswrapper[4492]: I0220 07:27:13.642285 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnw7z" event={"ID":"0311469a-95b9-4079-8a9f-e4d4a67e3b5f","Type":"ContainerStarted","Data":"da5d286835d69e8093f4aa691497d9549c69ff90bf505842ae077fd9b1d2f77a"} Feb 20 07:27:13 crc kubenswrapper[4492]: I0220 07:27:13.647502 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:27:14 crc kubenswrapper[4492]: I0220 07:27:14.653406 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnw7z" event={"ID":"0311469a-95b9-4079-8a9f-e4d4a67e3b5f","Type":"ContainerStarted","Data":"79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f"} Feb 20 07:27:16 crc kubenswrapper[4492]: I0220 07:27:16.704007 4492 generic.go:334] "Generic (PLEG): container finished" podID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerID="79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f" exitCode=0 Feb 20 07:27:16 crc kubenswrapper[4492]: I0220 07:27:16.704515 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnw7z" event={"ID":"0311469a-95b9-4079-8a9f-e4d4a67e3b5f","Type":"ContainerDied","Data":"79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f"} Feb 20 07:27:17 crc kubenswrapper[4492]: I0220 07:27:17.714066 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnw7z" event={"ID":"0311469a-95b9-4079-8a9f-e4d4a67e3b5f","Type":"ContainerStarted","Data":"e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364"} Feb 20 07:27:17 crc kubenswrapper[4492]: I0220 07:27:17.736757 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hnw7z" podStartSLOduration=3.173120177 podStartE2EDuration="6.736160811s" podCreationTimestamp="2026-02-20 07:27:11 +0000 UTC" firstStartedPulling="2026-02-20 07:27:13.64417511 +0000 UTC m=+2790.415464088" lastFinishedPulling="2026-02-20 07:27:17.207215744 +0000 UTC m=+2793.978504722" observedRunningTime="2026-02-20 07:27:17.729657879 +0000 UTC m=+2794.500946857" watchObservedRunningTime="2026-02-20 07:27:17.736160811 +0000 UTC m=+2794.507449790" Feb 20 07:27:22 crc kubenswrapper[4492]: I0220 07:27:22.109300 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:22 crc kubenswrapper[4492]: I0220 07:27:22.109822 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:23 crc kubenswrapper[4492]: I0220 07:27:23.187412 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hnw7z" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="registry-server" probeResult="failure" output=< Feb 20 07:27:23 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:27:23 crc kubenswrapper[4492]: > Feb 20 07:27:32 crc kubenswrapper[4492]: I0220 07:27:32.206319 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:32 crc kubenswrapper[4492]: I0220 07:27:32.334967 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:32 crc kubenswrapper[4492]: I0220 07:27:32.390752 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hnw7z"] Feb 20 07:27:33 crc kubenswrapper[4492]: I0220 07:27:33.856723 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hnw7z" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="registry-server" containerID="cri-o://e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364" gracePeriod=2 Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.675964 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.705969 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-catalog-content\") pod \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.706137 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7xzw\" (UniqueName: \"kubernetes.io/projected/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-kube-api-access-b7xzw\") pod \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.706315 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-utilities\") pod \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\" (UID: \"0311469a-95b9-4079-8a9f-e4d4a67e3b5f\") " Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.709576 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-utilities" (OuterVolumeSpecName: "utilities") pod "0311469a-95b9-4079-8a9f-e4d4a67e3b5f" (UID: "0311469a-95b9-4079-8a9f-e4d4a67e3b5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.723373 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-kube-api-access-b7xzw" (OuterVolumeSpecName: "kube-api-access-b7xzw") pod "0311469a-95b9-4079-8a9f-e4d4a67e3b5f" (UID: "0311469a-95b9-4079-8a9f-e4d4a67e3b5f"). InnerVolumeSpecName "kube-api-access-b7xzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.785697 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0311469a-95b9-4079-8a9f-e4d4a67e3b5f" (UID: "0311469a-95b9-4079-8a9f-e4d4a67e3b5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.808631 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7xzw\" (UniqueName: \"kubernetes.io/projected/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-kube-api-access-b7xzw\") on node \"crc\" DevicePath \"\"" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.808663 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.808673 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0311469a-95b9-4079-8a9f-e4d4a67e3b5f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.865399 4492 generic.go:334] "Generic (PLEG): container finished" podID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerID="e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364" exitCode=0 Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.865510 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hnw7z" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.865457 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnw7z" event={"ID":"0311469a-95b9-4079-8a9f-e4d4a67e3b5f","Type":"ContainerDied","Data":"e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364"} Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.865620 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hnw7z" event={"ID":"0311469a-95b9-4079-8a9f-e4d4a67e3b5f","Type":"ContainerDied","Data":"da5d286835d69e8093f4aa691497d9549c69ff90bf505842ae077fd9b1d2f77a"} Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.866516 4492 scope.go:117] "RemoveContainer" containerID="e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.897873 4492 scope.go:117] "RemoveContainer" containerID="79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.903431 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hnw7z"] Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.929203 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hnw7z"] Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.932181 4492 scope.go:117] "RemoveContainer" containerID="7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.953957 4492 scope.go:117] "RemoveContainer" containerID="e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364" Feb 20 07:27:34 crc kubenswrapper[4492]: E0220 07:27:34.956434 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364\": container with ID starting with e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364 not found: ID does not exist" containerID="e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.957035 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364"} err="failed to get container status \"e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364\": rpc error: code = NotFound desc = could not find container \"e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364\": container with ID starting with e935a5b041125713c673049003d86724764900ba68830c559e4c8a6af3072364 not found: ID does not exist" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.957065 4492 scope.go:117] "RemoveContainer" containerID="79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f" Feb 20 07:27:34 crc kubenswrapper[4492]: E0220 07:27:34.957323 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f\": container with ID starting with 79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f not found: ID does not exist" containerID="79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.957379 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f"} err="failed to get container status \"79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f\": rpc error: code = NotFound desc = could not find container \"79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f\": container with ID starting with 79833965f768a950a5dd4b87dcd8eea9420783665f6aed0ac1333c185185b68f not found: ID does not exist" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.957393 4492 scope.go:117] "RemoveContainer" containerID="7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6" Feb 20 07:27:34 crc kubenswrapper[4492]: E0220 07:27:34.957660 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6\": container with ID starting with 7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6 not found: ID does not exist" containerID="7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6" Feb 20 07:27:34 crc kubenswrapper[4492]: I0220 07:27:34.957683 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6"} err="failed to get container status \"7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6\": rpc error: code = NotFound desc = could not find container \"7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6\": container with ID starting with 7ca2eeaeadc446fa84796383787efd6572ce792092aa44b24391653a253f2ae6 not found: ID does not exist" Feb 20 07:27:35 crc kubenswrapper[4492]: I0220 07:27:35.566784 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" path="/var/lib/kubelet/pods/0311469a-95b9-4079-8a9f-e4d4a67e3b5f/volumes" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.047172 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xqg8q"] Feb 20 07:28:05 crc kubenswrapper[4492]: E0220 07:28:05.051751 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="registry-server" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.051778 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="registry-server" Feb 20 07:28:05 crc kubenswrapper[4492]: E0220 07:28:05.051808 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="extract-utilities" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.051815 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="extract-utilities" Feb 20 07:28:05 crc kubenswrapper[4492]: E0220 07:28:05.051835 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="extract-content" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.051843 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="extract-content" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.054312 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0311469a-95b9-4079-8a9f-e4d4a67e3b5f" containerName="registry-server" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.058839 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.097340 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzt8m\" (UniqueName: \"kubernetes.io/projected/f9143676-6114-4e83-b545-8c8b776e5889-kube-api-access-gzt8m\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.097653 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-utilities\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.097705 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-catalog-content\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.160432 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xqg8q"] Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.200099 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzt8m\" (UniqueName: \"kubernetes.io/projected/f9143676-6114-4e83-b545-8c8b776e5889-kube-api-access-gzt8m\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.200154 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-utilities\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.200276 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-catalog-content\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.203049 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-utilities\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.203662 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-catalog-content\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.243917 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzt8m\" (UniqueName: \"kubernetes.io/projected/f9143676-6114-4e83-b545-8c8b776e5889-kube-api-access-gzt8m\") pod \"community-operators-xqg8q\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:05 crc kubenswrapper[4492]: I0220 07:28:05.389601 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:06 crc kubenswrapper[4492]: I0220 07:28:06.166009 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xqg8q"] Feb 20 07:28:07 crc kubenswrapper[4492]: I0220 07:28:07.110418 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqg8q" event={"ID":"f9143676-6114-4e83-b545-8c8b776e5889","Type":"ContainerDied","Data":"050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1"} Feb 20 07:28:07 crc kubenswrapper[4492]: I0220 07:28:07.110545 4492 generic.go:334] "Generic (PLEG): container finished" podID="f9143676-6114-4e83-b545-8c8b776e5889" containerID="050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1" exitCode=0 Feb 20 07:28:07 crc kubenswrapper[4492]: I0220 07:28:07.111136 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqg8q" event={"ID":"f9143676-6114-4e83-b545-8c8b776e5889","Type":"ContainerStarted","Data":"6e789b20fe4e9451e871c3d5ea20f204cec293c79af49de2abc6af7583a354e4"} Feb 20 07:28:08 crc kubenswrapper[4492]: I0220 07:28:08.123347 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqg8q" event={"ID":"f9143676-6114-4e83-b545-8c8b776e5889","Type":"ContainerStarted","Data":"39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80"} Feb 20 07:28:09 crc kubenswrapper[4492]: I0220 07:28:09.134185 4492 generic.go:334] "Generic (PLEG): container finished" podID="f9143676-6114-4e83-b545-8c8b776e5889" containerID="39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80" exitCode=0 Feb 20 07:28:09 crc kubenswrapper[4492]: I0220 07:28:09.134259 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqg8q" event={"ID":"f9143676-6114-4e83-b545-8c8b776e5889","Type":"ContainerDied","Data":"39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80"} Feb 20 07:28:09 crc kubenswrapper[4492]: I0220 07:28:09.311198 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:28:09 crc kubenswrapper[4492]: I0220 07:28:09.312559 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:28:10 crc kubenswrapper[4492]: I0220 07:28:10.145139 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqg8q" event={"ID":"f9143676-6114-4e83-b545-8c8b776e5889","Type":"ContainerStarted","Data":"a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749"} Feb 20 07:28:10 crc kubenswrapper[4492]: I0220 07:28:10.169160 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xqg8q" podStartSLOduration=3.659253852 podStartE2EDuration="6.167727806s" podCreationTimestamp="2026-02-20 07:28:04 +0000 UTC" firstStartedPulling="2026-02-20 07:28:07.115176149 +0000 UTC m=+2843.886465116" lastFinishedPulling="2026-02-20 07:28:09.623650092 +0000 UTC m=+2846.394939070" observedRunningTime="2026-02-20 07:28:10.165216981 +0000 UTC m=+2846.936505959" watchObservedRunningTime="2026-02-20 07:28:10.167727806 +0000 UTC m=+2846.939016784" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.377715 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hrgzp"] Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.382569 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.429856 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrgzp"] Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.462844 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-catalog-content\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.462951 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-utilities\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.463190 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7c24\" (UniqueName: \"kubernetes.io/projected/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-kube-api-access-q7c24\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.566542 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-catalog-content\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.566715 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-utilities\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.566789 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7c24\" (UniqueName: \"kubernetes.io/projected/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-kube-api-access-q7c24\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.568411 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-catalog-content\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.568642 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-utilities\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.591160 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7c24\" (UniqueName: \"kubernetes.io/projected/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-kube-api-access-q7c24\") pod \"redhat-operators-hrgzp\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:12 crc kubenswrapper[4492]: I0220 07:28:12.731583 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:13 crc kubenswrapper[4492]: I0220 07:28:13.414797 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hrgzp"] Feb 20 07:28:14 crc kubenswrapper[4492]: I0220 07:28:14.179535 4492 generic.go:334] "Generic (PLEG): container finished" podID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerID="a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4" exitCode=0 Feb 20 07:28:14 crc kubenswrapper[4492]: I0220 07:28:14.179640 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrgzp" event={"ID":"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472","Type":"ContainerDied","Data":"a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4"} Feb 20 07:28:14 crc kubenswrapper[4492]: I0220 07:28:14.179887 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrgzp" event={"ID":"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472","Type":"ContainerStarted","Data":"8d2f9e4503f19286ae9876fe069817daab2e92392dfc141417a0e3471f52e767"} Feb 20 07:28:15 crc kubenswrapper[4492]: I0220 07:28:15.192090 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrgzp" event={"ID":"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472","Type":"ContainerStarted","Data":"04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9"} Feb 20 07:28:15 crc kubenswrapper[4492]: I0220 07:28:15.393060 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:15 crc kubenswrapper[4492]: I0220 07:28:15.393127 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:16 crc kubenswrapper[4492]: I0220 07:28:16.438878 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-xqg8q" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="registry-server" probeResult="failure" output=< Feb 20 07:28:16 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:28:16 crc kubenswrapper[4492]: > Feb 20 07:28:20 crc kubenswrapper[4492]: I0220 07:28:20.235352 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrgzp" event={"ID":"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472","Type":"ContainerDied","Data":"04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9"} Feb 20 07:28:20 crc kubenswrapper[4492]: I0220 07:28:20.236218 4492 generic.go:334] "Generic (PLEG): container finished" podID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerID="04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9" exitCode=0 Feb 20 07:28:21 crc kubenswrapper[4492]: I0220 07:28:21.245626 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrgzp" event={"ID":"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472","Type":"ContainerStarted","Data":"a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe"} Feb 20 07:28:21 crc kubenswrapper[4492]: I0220 07:28:21.318228 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hrgzp" podStartSLOduration=2.733412729 podStartE2EDuration="9.316222871s" podCreationTimestamp="2026-02-20 07:28:12 +0000 UTC" firstStartedPulling="2026-02-20 07:28:14.182255334 +0000 UTC m=+2850.953544313" lastFinishedPulling="2026-02-20 07:28:20.765065477 +0000 UTC m=+2857.536354455" observedRunningTime="2026-02-20 07:28:21.311145216 +0000 UTC m=+2858.082434194" watchObservedRunningTime="2026-02-20 07:28:21.316222871 +0000 UTC m=+2858.087511849" Feb 20 07:28:22 crc kubenswrapper[4492]: I0220 07:28:22.732086 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:22 crc kubenswrapper[4492]: I0220 07:28:22.732149 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:23 crc kubenswrapper[4492]: I0220 07:28:23.797132 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrgzp" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="registry-server" probeResult="failure" output=< Feb 20 07:28:23 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:28:23 crc kubenswrapper[4492]: > Feb 20 07:28:25 crc kubenswrapper[4492]: I0220 07:28:25.434800 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:25 crc kubenswrapper[4492]: I0220 07:28:25.474964 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:25 crc kubenswrapper[4492]: I0220 07:28:25.734942 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xqg8q"] Feb 20 07:28:27 crc kubenswrapper[4492]: I0220 07:28:27.293160 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xqg8q" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="registry-server" containerID="cri-o://a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749" gracePeriod=2 Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.248127 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.304682 4492 generic.go:334] "Generic (PLEG): container finished" podID="f9143676-6114-4e83-b545-8c8b776e5889" containerID="a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749" exitCode=0 Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.304946 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqg8q" event={"ID":"f9143676-6114-4e83-b545-8c8b776e5889","Type":"ContainerDied","Data":"a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749"} Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.304984 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqg8q" event={"ID":"f9143676-6114-4e83-b545-8c8b776e5889","Type":"ContainerDied","Data":"6e789b20fe4e9451e871c3d5ea20f204cec293c79af49de2abc6af7583a354e4"} Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.305062 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqg8q" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.306734 4492 scope.go:117] "RemoveContainer" containerID="a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.354128 4492 scope.go:117] "RemoveContainer" containerID="39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.385131 4492 scope.go:117] "RemoveContainer" containerID="050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.399489 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzt8m\" (UniqueName: \"kubernetes.io/projected/f9143676-6114-4e83-b545-8c8b776e5889-kube-api-access-gzt8m\") pod \"f9143676-6114-4e83-b545-8c8b776e5889\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.399545 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-catalog-content\") pod \"f9143676-6114-4e83-b545-8c8b776e5889\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.400859 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-utilities\") pod \"f9143676-6114-4e83-b545-8c8b776e5889\" (UID: \"f9143676-6114-4e83-b545-8c8b776e5889\") " Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.402575 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-utilities" (OuterVolumeSpecName: "utilities") pod "f9143676-6114-4e83-b545-8c8b776e5889" (UID: "f9143676-6114-4e83-b545-8c8b776e5889"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.416769 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9143676-6114-4e83-b545-8c8b776e5889-kube-api-access-gzt8m" (OuterVolumeSpecName: "kube-api-access-gzt8m") pod "f9143676-6114-4e83-b545-8c8b776e5889" (UID: "f9143676-6114-4e83-b545-8c8b776e5889"). InnerVolumeSpecName "kube-api-access-gzt8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.432497 4492 scope.go:117] "RemoveContainer" containerID="a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749" Feb 20 07:28:28 crc kubenswrapper[4492]: E0220 07:28:28.436287 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749\": container with ID starting with a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749 not found: ID does not exist" containerID="a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.436911 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749"} err="failed to get container status \"a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749\": rpc error: code = NotFound desc = could not find container \"a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749\": container with ID starting with a955021fe09e5b82f5301ee5b1a9451a341c728ddb7138056110c4f2bae84749 not found: ID does not exist" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.436956 4492 scope.go:117] "RemoveContainer" containerID="39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80" Feb 20 07:28:28 crc kubenswrapper[4492]: E0220 07:28:28.437427 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80\": container with ID starting with 39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80 not found: ID does not exist" containerID="39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.437449 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80"} err="failed to get container status \"39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80\": rpc error: code = NotFound desc = could not find container \"39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80\": container with ID starting with 39764b5edc9d1fa9a80ad0d8cb244d747f7f26e15fedb15a22f3ba258fa7ac80 not found: ID does not exist" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.437463 4492 scope.go:117] "RemoveContainer" containerID="050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1" Feb 20 07:28:28 crc kubenswrapper[4492]: E0220 07:28:28.437784 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1\": container with ID starting with 050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1 not found: ID does not exist" containerID="050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.437819 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1"} err="failed to get container status \"050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1\": rpc error: code = NotFound desc = could not find container \"050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1\": container with ID starting with 050c43aa94f4b72bc6f02a9cb88646194093c178723a945d2669d15cd3b0a8b1 not found: ID does not exist" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.478179 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9143676-6114-4e83-b545-8c8b776e5889" (UID: "f9143676-6114-4e83-b545-8c8b776e5889"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.503982 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.504009 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzt8m\" (UniqueName: \"kubernetes.io/projected/f9143676-6114-4e83-b545-8c8b776e5889-kube-api-access-gzt8m\") on node \"crc\" DevicePath \"\"" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.504020 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9143676-6114-4e83-b545-8c8b776e5889-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.650521 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xqg8q"] Feb 20 07:28:28 crc kubenswrapper[4492]: I0220 07:28:28.657010 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xqg8q"] Feb 20 07:28:28 crc kubenswrapper[4492]: E0220 07:28:28.759877 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9143676_6114_4e83_b545_8c8b776e5889.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9143676_6114_4e83_b545_8c8b776e5889.slice/crio-6e789b20fe4e9451e871c3d5ea20f204cec293c79af49de2abc6af7583a354e4\": RecentStats: unable to find data in memory cache]" Feb 20 07:28:29 crc kubenswrapper[4492]: I0220 07:28:29.568679 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9143676-6114-4e83-b545-8c8b776e5889" path="/var/lib/kubelet/pods/f9143676-6114-4e83-b545-8c8b776e5889/volumes" Feb 20 07:28:33 crc kubenswrapper[4492]: I0220 07:28:33.771099 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hrgzp" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="registry-server" probeResult="failure" output=< Feb 20 07:28:33 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:28:33 crc kubenswrapper[4492]: > Feb 20 07:28:39 crc kubenswrapper[4492]: I0220 07:28:39.312245 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:28:39 crc kubenswrapper[4492]: I0220 07:28:39.313314 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:28:42 crc kubenswrapper[4492]: I0220 07:28:42.778500 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:42 crc kubenswrapper[4492]: I0220 07:28:42.815356 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:43 crc kubenswrapper[4492]: I0220 07:28:43.154955 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrgzp"] Feb 20 07:28:44 crc kubenswrapper[4492]: I0220 07:28:44.429057 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hrgzp" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="registry-server" containerID="cri-o://a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe" gracePeriod=2 Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.131322 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.313528 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-utilities\") pod \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.313922 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-catalog-content\") pod \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.313979 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7c24\" (UniqueName: \"kubernetes.io/projected/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-kube-api-access-q7c24\") pod \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\" (UID: \"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472\") " Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.316075 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-utilities" (OuterVolumeSpecName: "utilities") pod "84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" (UID: "84f3d861-5c6c-4ba0-88e6-ec76e0cc8472"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.327684 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-kube-api-access-q7c24" (OuterVolumeSpecName: "kube-api-access-q7c24") pod "84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" (UID: "84f3d861-5c6c-4ba0-88e6-ec76e0cc8472"). InnerVolumeSpecName "kube-api-access-q7c24". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.408264 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" (UID: "84f3d861-5c6c-4ba0-88e6-ec76e0cc8472"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.417036 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.417065 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7c24\" (UniqueName: \"kubernetes.io/projected/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-kube-api-access-q7c24\") on node \"crc\" DevicePath \"\"" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.417078 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.439427 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrgzp" event={"ID":"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472","Type":"ContainerDied","Data":"a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe"} Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.439517 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hrgzp" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.439450 4492 generic.go:334] "Generic (PLEG): container finished" podID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerID="a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe" exitCode=0 Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.439728 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hrgzp" event={"ID":"84f3d861-5c6c-4ba0-88e6-ec76e0cc8472","Type":"ContainerDied","Data":"8d2f9e4503f19286ae9876fe069817daab2e92392dfc141417a0e3471f52e767"} Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.439899 4492 scope.go:117] "RemoveContainer" containerID="a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.477273 4492 scope.go:117] "RemoveContainer" containerID="04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.478517 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hrgzp"] Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.485084 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hrgzp"] Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.495708 4492 scope.go:117] "RemoveContainer" containerID="a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.530681 4492 scope.go:117] "RemoveContainer" containerID="a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe" Feb 20 07:28:45 crc kubenswrapper[4492]: E0220 07:28:45.533987 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe\": container with ID starting with a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe not found: ID does not exist" containerID="a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.534057 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe"} err="failed to get container status \"a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe\": rpc error: code = NotFound desc = could not find container \"a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe\": container with ID starting with a995ec59a2e7925b28f695e2f0122dbe7f6491b6ae0f4328df0d9d5af2d334fe not found: ID does not exist" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.534079 4492 scope.go:117] "RemoveContainer" containerID="04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9" Feb 20 07:28:45 crc kubenswrapper[4492]: E0220 07:28:45.534417 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9\": container with ID starting with 04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9 not found: ID does not exist" containerID="04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.534456 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9"} err="failed to get container status \"04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9\": rpc error: code = NotFound desc = could not find container \"04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9\": container with ID starting with 04cc386aa9a538f28e45dbfc381b74935ef4c75dffbc1b34d084335ef28c69d9 not found: ID does not exist" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.534510 4492 scope.go:117] "RemoveContainer" containerID="a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4" Feb 20 07:28:45 crc kubenswrapper[4492]: E0220 07:28:45.534927 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4\": container with ID starting with a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4 not found: ID does not exist" containerID="a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.534967 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4"} err="failed to get container status \"a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4\": rpc error: code = NotFound desc = could not find container \"a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4\": container with ID starting with a8097a2131c55b94b377e44343b61246becde7f6c198b58145828478ebe689c4 not found: ID does not exist" Feb 20 07:28:45 crc kubenswrapper[4492]: I0220 07:28:45.566688 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" path="/var/lib/kubelet/pods/84f3d861-5c6c-4ba0-88e6-ec76e0cc8472/volumes" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.295911 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-whjn9"] Feb 20 07:29:05 crc kubenswrapper[4492]: E0220 07:29:05.297346 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="extract-content" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.297373 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="extract-content" Feb 20 07:29:05 crc kubenswrapper[4492]: E0220 07:29:05.297405 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="extract-utilities" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.297411 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="extract-utilities" Feb 20 07:29:05 crc kubenswrapper[4492]: E0220 07:29:05.297425 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="extract-content" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.297431 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="extract-content" Feb 20 07:29:05 crc kubenswrapper[4492]: E0220 07:29:05.297441 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="registry-server" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.297446 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="registry-server" Feb 20 07:29:05 crc kubenswrapper[4492]: E0220 07:29:05.297456 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="registry-server" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.297461 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="registry-server" Feb 20 07:29:05 crc kubenswrapper[4492]: E0220 07:29:05.297522 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="extract-utilities" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.297529 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="extract-utilities" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.298392 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="84f3d861-5c6c-4ba0-88e6-ec76e0cc8472" containerName="registry-server" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.298451 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9143676-6114-4e83-b545-8c8b776e5889" containerName="registry-server" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.302787 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.323139 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-whjn9"] Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.354001 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjh6z\" (UniqueName: \"kubernetes.io/projected/94028668-7a67-49e5-8838-de3ae88f16e5-kube-api-access-fjh6z\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.354227 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-utilities\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.354273 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-catalog-content\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.456558 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-utilities\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.456597 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-catalog-content\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.456670 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjh6z\" (UniqueName: \"kubernetes.io/projected/94028668-7a67-49e5-8838-de3ae88f16e5-kube-api-access-fjh6z\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.459872 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-utilities\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.460533 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-catalog-content\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.486030 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjh6z\" (UniqueName: \"kubernetes.io/projected/94028668-7a67-49e5-8838-de3ae88f16e5-kube-api-access-fjh6z\") pod \"redhat-marketplace-whjn9\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:05 crc kubenswrapper[4492]: I0220 07:29:05.622466 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:06 crc kubenswrapper[4492]: I0220 07:29:06.170249 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-whjn9"] Feb 20 07:29:06 crc kubenswrapper[4492]: I0220 07:29:06.589000 4492 generic.go:334] "Generic (PLEG): container finished" podID="94028668-7a67-49e5-8838-de3ae88f16e5" containerID="f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49" exitCode=0 Feb 20 07:29:06 crc kubenswrapper[4492]: I0220 07:29:06.589099 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whjn9" event={"ID":"94028668-7a67-49e5-8838-de3ae88f16e5","Type":"ContainerDied","Data":"f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49"} Feb 20 07:29:06 crc kubenswrapper[4492]: I0220 07:29:06.589252 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whjn9" event={"ID":"94028668-7a67-49e5-8838-de3ae88f16e5","Type":"ContainerStarted","Data":"c9068579f230907fd737210b04baa7adf2cead97d66627374f19477acc12d120"} Feb 20 07:29:07 crc kubenswrapper[4492]: I0220 07:29:07.599330 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whjn9" event={"ID":"94028668-7a67-49e5-8838-de3ae88f16e5","Type":"ContainerStarted","Data":"586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7"} Feb 20 07:29:08 crc kubenswrapper[4492]: I0220 07:29:08.608225 4492 generic.go:334] "Generic (PLEG): container finished" podID="94028668-7a67-49e5-8838-de3ae88f16e5" containerID="586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7" exitCode=0 Feb 20 07:29:08 crc kubenswrapper[4492]: I0220 07:29:08.608499 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whjn9" event={"ID":"94028668-7a67-49e5-8838-de3ae88f16e5","Type":"ContainerDied","Data":"586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7"} Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.312553 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.313709 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.313892 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.315309 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.315707 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" gracePeriod=600 Feb 20 07:29:09 crc kubenswrapper[4492]: E0220 07:29:09.449589 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.618884 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whjn9" event={"ID":"94028668-7a67-49e5-8838-de3ae88f16e5","Type":"ContainerStarted","Data":"2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c"} Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.622000 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" exitCode=0 Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.622047 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9"} Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.622418 4492 scope.go:117] "RemoveContainer" containerID="ea2dd75e96a336a1f60c20e23825e1bcd69e3e675203868cf884eb3d626bc2c4" Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.622589 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:29:09 crc kubenswrapper[4492]: E0220 07:29:09.622960 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:29:09 crc kubenswrapper[4492]: I0220 07:29:09.637163 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-whjn9" podStartSLOduration=2.079780912 podStartE2EDuration="4.637148164s" podCreationTimestamp="2026-02-20 07:29:05 +0000 UTC" firstStartedPulling="2026-02-20 07:29:06.590599928 +0000 UTC m=+2903.361888905" lastFinishedPulling="2026-02-20 07:29:09.147967179 +0000 UTC m=+2905.919256157" observedRunningTime="2026-02-20 07:29:09.633899968 +0000 UTC m=+2906.405188946" watchObservedRunningTime="2026-02-20 07:29:09.637148164 +0000 UTC m=+2906.408437142" Feb 20 07:29:15 crc kubenswrapper[4492]: I0220 07:29:15.623063 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:15 crc kubenswrapper[4492]: I0220 07:29:15.624787 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:15 crc kubenswrapper[4492]: I0220 07:29:15.662602 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:15 crc kubenswrapper[4492]: I0220 07:29:15.714701 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:15 crc kubenswrapper[4492]: I0220 07:29:15.911956 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-whjn9"] Feb 20 07:29:17 crc kubenswrapper[4492]: I0220 07:29:17.692636 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-whjn9" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="registry-server" containerID="cri-o://2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c" gracePeriod=2 Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.438505 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.616264 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-catalog-content\") pod \"94028668-7a67-49e5-8838-de3ae88f16e5\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.616450 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjh6z\" (UniqueName: \"kubernetes.io/projected/94028668-7a67-49e5-8838-de3ae88f16e5-kube-api-access-fjh6z\") pod \"94028668-7a67-49e5-8838-de3ae88f16e5\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.616574 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-utilities\") pod \"94028668-7a67-49e5-8838-de3ae88f16e5\" (UID: \"94028668-7a67-49e5-8838-de3ae88f16e5\") " Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.617566 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-utilities" (OuterVolumeSpecName: "utilities") pod "94028668-7a67-49e5-8838-de3ae88f16e5" (UID: "94028668-7a67-49e5-8838-de3ae88f16e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.638358 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94028668-7a67-49e5-8838-de3ae88f16e5-kube-api-access-fjh6z" (OuterVolumeSpecName: "kube-api-access-fjh6z") pod "94028668-7a67-49e5-8838-de3ae88f16e5" (UID: "94028668-7a67-49e5-8838-de3ae88f16e5"). InnerVolumeSpecName "kube-api-access-fjh6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.650867 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94028668-7a67-49e5-8838-de3ae88f16e5" (UID: "94028668-7a67-49e5-8838-de3ae88f16e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.700023 4492 generic.go:334] "Generic (PLEG): container finished" podID="94028668-7a67-49e5-8838-de3ae88f16e5" containerID="2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c" exitCode=0 Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.700063 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whjn9" event={"ID":"94028668-7a67-49e5-8838-de3ae88f16e5","Type":"ContainerDied","Data":"2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c"} Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.700092 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whjn9" event={"ID":"94028668-7a67-49e5-8838-de3ae88f16e5","Type":"ContainerDied","Data":"c9068579f230907fd737210b04baa7adf2cead97d66627374f19477acc12d120"} Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.700110 4492 scope.go:117] "RemoveContainer" containerID="2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.700242 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whjn9" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.720363 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.720392 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjh6z\" (UniqueName: \"kubernetes.io/projected/94028668-7a67-49e5-8838-de3ae88f16e5-kube-api-access-fjh6z\") on node \"crc\" DevicePath \"\"" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.720402 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94028668-7a67-49e5-8838-de3ae88f16e5-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.729920 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-whjn9"] Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.735166 4492 scope.go:117] "RemoveContainer" containerID="586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.739057 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-whjn9"] Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.762782 4492 scope.go:117] "RemoveContainer" containerID="f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.792727 4492 scope.go:117] "RemoveContainer" containerID="2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c" Feb 20 07:29:18 crc kubenswrapper[4492]: E0220 07:29:18.795841 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c\": container with ID starting with 2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c not found: ID does not exist" containerID="2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.796531 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c"} err="failed to get container status \"2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c\": rpc error: code = NotFound desc = could not find container \"2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c\": container with ID starting with 2534b8ea209a01d2d40d4008c789cf054117495bde6fa633db46b702d2aa7b9c not found: ID does not exist" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.796622 4492 scope.go:117] "RemoveContainer" containerID="586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7" Feb 20 07:29:18 crc kubenswrapper[4492]: E0220 07:29:18.796993 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7\": container with ID starting with 586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7 not found: ID does not exist" containerID="586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.797072 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7"} err="failed to get container status \"586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7\": rpc error: code = NotFound desc = could not find container \"586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7\": container with ID starting with 586b2b7f040c84537d8de2457cfc6877b7634d5004d0f166c156578bac14baf7 not found: ID does not exist" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.797133 4492 scope.go:117] "RemoveContainer" containerID="f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49" Feb 20 07:29:18 crc kubenswrapper[4492]: E0220 07:29:18.797503 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49\": container with ID starting with f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49 not found: ID does not exist" containerID="f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49" Feb 20 07:29:18 crc kubenswrapper[4492]: I0220 07:29:18.797584 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49"} err="failed to get container status \"f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49\": rpc error: code = NotFound desc = could not find container \"f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49\": container with ID starting with f8b7d149b7367b7f640e6583a675f0b5a221bf97c96724d84a97d5e0305a2f49 not found: ID does not exist" Feb 20 07:29:19 crc kubenswrapper[4492]: I0220 07:29:19.566146 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" path="/var/lib/kubelet/pods/94028668-7a67-49e5-8838-de3ae88f16e5/volumes" Feb 20 07:29:25 crc kubenswrapper[4492]: I0220 07:29:25.556612 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:29:25 crc kubenswrapper[4492]: E0220 07:29:25.557332 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:29:36 crc kubenswrapper[4492]: I0220 07:29:36.557632 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:29:36 crc kubenswrapper[4492]: E0220 07:29:36.560040 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:29:50 crc kubenswrapper[4492]: I0220 07:29:50.557162 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:29:50 crc kubenswrapper[4492]: E0220 07:29:50.558083 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.632541 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm"] Feb 20 07:30:00 crc kubenswrapper[4492]: E0220 07:30:00.637918 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="extract-utilities" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.637943 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="extract-utilities" Feb 20 07:30:00 crc kubenswrapper[4492]: E0220 07:30:00.638149 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="registry-server" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.638159 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="registry-server" Feb 20 07:30:00 crc kubenswrapper[4492]: E0220 07:30:00.638180 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="extract-content" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.638192 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="extract-content" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.639142 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="94028668-7a67-49e5-8838-de3ae88f16e5" containerName="registry-server" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.643527 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.671448 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j6m2\" (UniqueName: \"kubernetes.io/projected/a01f19eb-c60f-4612-8a50-a78564ab6654-kube-api-access-2j6m2\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.671537 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a01f19eb-c60f-4612-8a50-a78564ab6654-secret-volume\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.671620 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f19eb-c60f-4612-8a50-a78564ab6654-config-volume\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.693944 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.694689 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.744676 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm"] Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.772866 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j6m2\" (UniqueName: \"kubernetes.io/projected/a01f19eb-c60f-4612-8a50-a78564ab6654-kube-api-access-2j6m2\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.773194 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a01f19eb-c60f-4612-8a50-a78564ab6654-secret-volume\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.773263 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f19eb-c60f-4612-8a50-a78564ab6654-config-volume\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.782708 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f19eb-c60f-4612-8a50-a78564ab6654-config-volume\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.791644 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a01f19eb-c60f-4612-8a50-a78564ab6654-secret-volume\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.792163 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j6m2\" (UniqueName: \"kubernetes.io/projected/a01f19eb-c60f-4612-8a50-a78564ab6654-kube-api-access-2j6m2\") pod \"collect-profiles-29526210-mk4qm\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:00 crc kubenswrapper[4492]: I0220 07:30:00.970190 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:01 crc kubenswrapper[4492]: I0220 07:30:01.782876 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm"] Feb 20 07:30:02 crc kubenswrapper[4492]: I0220 07:30:02.052574 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" event={"ID":"a01f19eb-c60f-4612-8a50-a78564ab6654","Type":"ContainerStarted","Data":"8eee68dfa83451ae4b028554f3a937821af25e2d4c060d765336df54f980093e"} Feb 20 07:30:02 crc kubenswrapper[4492]: I0220 07:30:02.052617 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" event={"ID":"a01f19eb-c60f-4612-8a50-a78564ab6654","Type":"ContainerStarted","Data":"f05437827ed5eeaab95e509c3a7aa9fc8a4f6d07f3b015beb2c9e9f3fa3287e0"} Feb 20 07:30:02 crc kubenswrapper[4492]: I0220 07:30:02.072122 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" podStartSLOduration=2.071059497 podStartE2EDuration="2.071059497s" podCreationTimestamp="2026-02-20 07:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 07:30:02.068747216 +0000 UTC m=+2958.840036194" watchObservedRunningTime="2026-02-20 07:30:02.071059497 +0000 UTC m=+2958.842348475" Feb 20 07:30:03 crc kubenswrapper[4492]: I0220 07:30:03.064072 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" event={"ID":"a01f19eb-c60f-4612-8a50-a78564ab6654","Type":"ContainerDied","Data":"8eee68dfa83451ae4b028554f3a937821af25e2d4c060d765336df54f980093e"} Feb 20 07:30:03 crc kubenswrapper[4492]: I0220 07:30:03.065370 4492 generic.go:334] "Generic (PLEG): container finished" podID="a01f19eb-c60f-4612-8a50-a78564ab6654" containerID="8eee68dfa83451ae4b028554f3a937821af25e2d4c060d765336df54f980093e" exitCode=0 Feb 20 07:30:03 crc kubenswrapper[4492]: I0220 07:30:03.568905 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:30:03 crc kubenswrapper[4492]: E0220 07:30:03.570180 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.683166 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.847529 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j6m2\" (UniqueName: \"kubernetes.io/projected/a01f19eb-c60f-4612-8a50-a78564ab6654-kube-api-access-2j6m2\") pod \"a01f19eb-c60f-4612-8a50-a78564ab6654\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.847681 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a01f19eb-c60f-4612-8a50-a78564ab6654-secret-volume\") pod \"a01f19eb-c60f-4612-8a50-a78564ab6654\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.847714 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f19eb-c60f-4612-8a50-a78564ab6654-config-volume\") pod \"a01f19eb-c60f-4612-8a50-a78564ab6654\" (UID: \"a01f19eb-c60f-4612-8a50-a78564ab6654\") " Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.851073 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01f19eb-c60f-4612-8a50-a78564ab6654-config-volume" (OuterVolumeSpecName: "config-volume") pod "a01f19eb-c60f-4612-8a50-a78564ab6654" (UID: "a01f19eb-c60f-4612-8a50-a78564ab6654"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.864557 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a01f19eb-c60f-4612-8a50-a78564ab6654-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a01f19eb-c60f-4612-8a50-a78564ab6654" (UID: "a01f19eb-c60f-4612-8a50-a78564ab6654"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.867184 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01f19eb-c60f-4612-8a50-a78564ab6654-kube-api-access-2j6m2" (OuterVolumeSpecName: "kube-api-access-2j6m2") pod "a01f19eb-c60f-4612-8a50-a78564ab6654" (UID: "a01f19eb-c60f-4612-8a50-a78564ab6654"). InnerVolumeSpecName "kube-api-access-2j6m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.884295 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv"] Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.890001 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526165-7vskv"] Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.949981 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a01f19eb-c60f-4612-8a50-a78564ab6654-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.950008 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f19eb-c60f-4612-8a50-a78564ab6654-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:30:04 crc kubenswrapper[4492]: I0220 07:30:04.950028 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j6m2\" (UniqueName: \"kubernetes.io/projected/a01f19eb-c60f-4612-8a50-a78564ab6654-kube-api-access-2j6m2\") on node \"crc\" DevicePath \"\"" Feb 20 07:30:05 crc kubenswrapper[4492]: I0220 07:30:05.083745 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" event={"ID":"a01f19eb-c60f-4612-8a50-a78564ab6654","Type":"ContainerDied","Data":"f05437827ed5eeaab95e509c3a7aa9fc8a4f6d07f3b015beb2c9e9f3fa3287e0"} Feb 20 07:30:05 crc kubenswrapper[4492]: I0220 07:30:05.084250 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm" Feb 20 07:30:05 crc kubenswrapper[4492]: I0220 07:30:05.084383 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f05437827ed5eeaab95e509c3a7aa9fc8a4f6d07f3b015beb2c9e9f3fa3287e0" Feb 20 07:30:05 crc kubenswrapper[4492]: I0220 07:30:05.567160 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="284c2ed2-e161-4e3b-9ceb-891c71dc4e1f" path="/var/lib/kubelet/pods/284c2ed2-e161-4e3b-9ceb-891c71dc4e1f/volumes" Feb 20 07:30:07 crc kubenswrapper[4492]: I0220 07:30:07.247002 4492 scope.go:117] "RemoveContainer" containerID="db8b5a49afb523ad43b8c4093110a8b61adbac549e7df2fe0a92fa35ebeef35a" Feb 20 07:30:14 crc kubenswrapper[4492]: I0220 07:30:14.557280 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:30:14 crc kubenswrapper[4492]: E0220 07:30:14.559232 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:30:27 crc kubenswrapper[4492]: I0220 07:30:27.558055 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:30:27 crc kubenswrapper[4492]: E0220 07:30:27.558834 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:30:39 crc kubenswrapper[4492]: I0220 07:30:39.557985 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:30:39 crc kubenswrapper[4492]: E0220 07:30:39.558975 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:30:51 crc kubenswrapper[4492]: I0220 07:30:51.557862 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:30:51 crc kubenswrapper[4492]: E0220 07:30:51.558622 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:31:02 crc kubenswrapper[4492]: I0220 07:31:02.556573 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:31:02 crc kubenswrapper[4492]: E0220 07:31:02.557156 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:31:17 crc kubenswrapper[4492]: I0220 07:31:17.557631 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:31:17 crc kubenswrapper[4492]: E0220 07:31:17.558548 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:31:29 crc kubenswrapper[4492]: I0220 07:31:29.557460 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:31:29 crc kubenswrapper[4492]: E0220 07:31:29.559098 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:31:40 crc kubenswrapper[4492]: I0220 07:31:40.557451 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:31:40 crc kubenswrapper[4492]: E0220 07:31:40.558408 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:31:47 crc kubenswrapper[4492]: E0220 07:31:47.984081 4492 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.25:53998->192.168.26.25:44007: write tcp 192.168.26.25:53998->192.168.26.25:44007: write: broken pipe Feb 20 07:31:53 crc kubenswrapper[4492]: I0220 07:31:53.562938 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:31:53 crc kubenswrapper[4492]: E0220 07:31:53.564612 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:32:05 crc kubenswrapper[4492]: I0220 07:32:05.556825 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:32:05 crc kubenswrapper[4492]: E0220 07:32:05.558792 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:32:19 crc kubenswrapper[4492]: I0220 07:32:19.556933 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:32:19 crc kubenswrapper[4492]: E0220 07:32:19.557679 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:32:32 crc kubenswrapper[4492]: I0220 07:32:32.556693 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:32:32 crc kubenswrapper[4492]: E0220 07:32:32.557345 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:32:44 crc kubenswrapper[4492]: I0220 07:32:44.556961 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:32:44 crc kubenswrapper[4492]: E0220 07:32:44.557749 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:32:59 crc kubenswrapper[4492]: I0220 07:32:59.557881 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:32:59 crc kubenswrapper[4492]: E0220 07:32:59.558756 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:33:12 crc kubenswrapper[4492]: I0220 07:33:12.557573 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:33:12 crc kubenswrapper[4492]: E0220 07:33:12.558271 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:33:25 crc kubenswrapper[4492]: I0220 07:33:25.556637 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:33:25 crc kubenswrapper[4492]: E0220 07:33:25.557257 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:33:38 crc kubenswrapper[4492]: I0220 07:33:38.557003 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:33:38 crc kubenswrapper[4492]: E0220 07:33:38.557777 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:33:50 crc kubenswrapper[4492]: I0220 07:33:50.557060 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:33:50 crc kubenswrapper[4492]: E0220 07:33:50.558886 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:34:01 crc kubenswrapper[4492]: I0220 07:34:01.557869 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:34:01 crc kubenswrapper[4492]: E0220 07:34:01.558452 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:34:13 crc kubenswrapper[4492]: I0220 07:34:13.561633 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:34:13 crc kubenswrapper[4492]: I0220 07:34:13.978461 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"172222d62e8e3557fb3aeaadcdb3793eab777225daeaf180018aac17f6dcf60d"} Feb 20 07:36:39 crc kubenswrapper[4492]: I0220 07:36:39.314634 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:36:39 crc kubenswrapper[4492]: I0220 07:36:39.317576 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:37:09 crc kubenswrapper[4492]: I0220 07:37:09.311609 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:37:09 crc kubenswrapper[4492]: I0220 07:37:09.311980 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:37:39 crc kubenswrapper[4492]: I0220 07:37:39.311430 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:37:39 crc kubenswrapper[4492]: I0220 07:37:39.312083 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:37:39 crc kubenswrapper[4492]: I0220 07:37:39.312833 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:37:39 crc kubenswrapper[4492]: I0220 07:37:39.315733 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"172222d62e8e3557fb3aeaadcdb3793eab777225daeaf180018aac17f6dcf60d"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:37:39 crc kubenswrapper[4492]: I0220 07:37:39.317008 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://172222d62e8e3557fb3aeaadcdb3793eab777225daeaf180018aac17f6dcf60d" gracePeriod=600 Feb 20 07:37:40 crc kubenswrapper[4492]: I0220 07:37:40.434975 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="172222d62e8e3557fb3aeaadcdb3793eab777225daeaf180018aac17f6dcf60d" exitCode=0 Feb 20 07:37:40 crc kubenswrapper[4492]: I0220 07:37:40.435305 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"172222d62e8e3557fb3aeaadcdb3793eab777225daeaf180018aac17f6dcf60d"} Feb 20 07:37:40 crc kubenswrapper[4492]: I0220 07:37:40.436251 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165"} Feb 20 07:37:40 crc kubenswrapper[4492]: I0220 07:37:40.437301 4492 scope.go:117] "RemoveContainer" containerID="8c3b6b0b8ca1931f021fc7dd624f4720ec231a33c08e4776993ce562246948b9" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.759602 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lghp6"] Feb 20 07:37:42 crc kubenswrapper[4492]: E0220 07:37:42.763248 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01f19eb-c60f-4612-8a50-a78564ab6654" containerName="collect-profiles" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.763335 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01f19eb-c60f-4612-8a50-a78564ab6654" containerName="collect-profiles" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.768839 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01f19eb-c60f-4612-8a50-a78564ab6654" containerName="collect-profiles" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.774610 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.817843 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8fnv\" (UniqueName: \"kubernetes.io/projected/3fd6c20f-0afb-48db-b362-724d3ff70682-kube-api-access-x8fnv\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.817985 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-catalog-content\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.818107 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-utilities\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.824549 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lghp6"] Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.919504 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-utilities\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.919758 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8fnv\" (UniqueName: \"kubernetes.io/projected/3fd6c20f-0afb-48db-b362-724d3ff70682-kube-api-access-x8fnv\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.919841 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-catalog-content\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.922644 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-utilities\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.923180 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-catalog-content\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:42 crc kubenswrapper[4492]: I0220 07:37:42.949311 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8fnv\" (UniqueName: \"kubernetes.io/projected/3fd6c20f-0afb-48db-b362-724d3ff70682-kube-api-access-x8fnv\") pod \"certified-operators-lghp6\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:43 crc kubenswrapper[4492]: I0220 07:37:43.104664 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:43 crc kubenswrapper[4492]: I0220 07:37:43.894932 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lghp6"] Feb 20 07:37:44 crc kubenswrapper[4492]: I0220 07:37:44.477760 4492 generic.go:334] "Generic (PLEG): container finished" podID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerID="f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74" exitCode=0 Feb 20 07:37:44 crc kubenswrapper[4492]: I0220 07:37:44.478034 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lghp6" event={"ID":"3fd6c20f-0afb-48db-b362-724d3ff70682","Type":"ContainerDied","Data":"f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74"} Feb 20 07:37:44 crc kubenswrapper[4492]: I0220 07:37:44.478065 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lghp6" event={"ID":"3fd6c20f-0afb-48db-b362-724d3ff70682","Type":"ContainerStarted","Data":"8fdd9c9178dda94a0161c05559a8316d525f302521d7b588c73155ab83c1b3ea"} Feb 20 07:37:44 crc kubenswrapper[4492]: I0220 07:37:44.483108 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:37:45 crc kubenswrapper[4492]: I0220 07:37:45.513079 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lghp6" event={"ID":"3fd6c20f-0afb-48db-b362-724d3ff70682","Type":"ContainerStarted","Data":"1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554"} Feb 20 07:37:46 crc kubenswrapper[4492]: I0220 07:37:46.527404 4492 generic.go:334] "Generic (PLEG): container finished" podID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerID="1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554" exitCode=0 Feb 20 07:37:46 crc kubenswrapper[4492]: I0220 07:37:46.527724 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lghp6" event={"ID":"3fd6c20f-0afb-48db-b362-724d3ff70682","Type":"ContainerDied","Data":"1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554"} Feb 20 07:37:47 crc kubenswrapper[4492]: I0220 07:37:47.537110 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lghp6" event={"ID":"3fd6c20f-0afb-48db-b362-724d3ff70682","Type":"ContainerStarted","Data":"f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f"} Feb 20 07:37:47 crc kubenswrapper[4492]: I0220 07:37:47.556623 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lghp6" podStartSLOduration=3.055455123 podStartE2EDuration="5.555523988s" podCreationTimestamp="2026-02-20 07:37:42 +0000 UTC" firstStartedPulling="2026-02-20 07:37:44.480340986 +0000 UTC m=+3421.251629964" lastFinishedPulling="2026-02-20 07:37:46.980409851 +0000 UTC m=+3423.751698829" observedRunningTime="2026-02-20 07:37:47.554566341 +0000 UTC m=+3424.325855320" watchObservedRunningTime="2026-02-20 07:37:47.555523988 +0000 UTC m=+3424.326812966" Feb 20 07:37:53 crc kubenswrapper[4492]: I0220 07:37:53.105928 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:53 crc kubenswrapper[4492]: I0220 07:37:53.106559 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:37:54 crc kubenswrapper[4492]: I0220 07:37:54.151025 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lghp6" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="registry-server" probeResult="failure" output=< Feb 20 07:37:54 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:37:54 crc kubenswrapper[4492]: > Feb 20 07:38:03 crc kubenswrapper[4492]: I0220 07:38:03.154637 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:38:03 crc kubenswrapper[4492]: I0220 07:38:03.191761 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:38:03 crc kubenswrapper[4492]: I0220 07:38:03.239923 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lghp6"] Feb 20 07:38:04 crc kubenswrapper[4492]: I0220 07:38:04.667982 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lghp6" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="registry-server" containerID="cri-o://f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f" gracePeriod=2 Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.384129 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.438565 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8fnv\" (UniqueName: \"kubernetes.io/projected/3fd6c20f-0afb-48db-b362-724d3ff70682-kube-api-access-x8fnv\") pod \"3fd6c20f-0afb-48db-b362-724d3ff70682\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.438633 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-catalog-content\") pod \"3fd6c20f-0afb-48db-b362-724d3ff70682\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.438654 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-utilities\") pod \"3fd6c20f-0afb-48db-b362-724d3ff70682\" (UID: \"3fd6c20f-0afb-48db-b362-724d3ff70682\") " Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.439775 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-utilities" (OuterVolumeSpecName: "utilities") pod "3fd6c20f-0afb-48db-b362-724d3ff70682" (UID: "3fd6c20f-0afb-48db-b362-724d3ff70682"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.452818 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fd6c20f-0afb-48db-b362-724d3ff70682-kube-api-access-x8fnv" (OuterVolumeSpecName: "kube-api-access-x8fnv") pod "3fd6c20f-0afb-48db-b362-724d3ff70682" (UID: "3fd6c20f-0afb-48db-b362-724d3ff70682"). InnerVolumeSpecName "kube-api-access-x8fnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.516122 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fd6c20f-0afb-48db-b362-724d3ff70682" (UID: "3fd6c20f-0afb-48db-b362-724d3ff70682"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.540817 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8fnv\" (UniqueName: \"kubernetes.io/projected/3fd6c20f-0afb-48db-b362-724d3ff70682-kube-api-access-x8fnv\") on node \"crc\" DevicePath \"\"" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.540847 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.540858 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c20f-0afb-48db-b362-724d3ff70682-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.679034 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lghp6" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.679070 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lghp6" event={"ID":"3fd6c20f-0afb-48db-b362-724d3ff70682","Type":"ContainerDied","Data":"f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f"} Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.679781 4492 scope.go:117] "RemoveContainer" containerID="f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.679738 4492 generic.go:334] "Generic (PLEG): container finished" podID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerID="f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f" exitCode=0 Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.680674 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lghp6" event={"ID":"3fd6c20f-0afb-48db-b362-724d3ff70682","Type":"ContainerDied","Data":"8fdd9c9178dda94a0161c05559a8316d525f302521d7b588c73155ab83c1b3ea"} Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.709547 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lghp6"] Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.732456 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lghp6"] Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.746622 4492 scope.go:117] "RemoveContainer" containerID="1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.777203 4492 scope.go:117] "RemoveContainer" containerID="f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.806254 4492 scope.go:117] "RemoveContainer" containerID="f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f" Feb 20 07:38:05 crc kubenswrapper[4492]: E0220 07:38:05.809323 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f\": container with ID starting with f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f not found: ID does not exist" containerID="f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.811102 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f"} err="failed to get container status \"f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f\": rpc error: code = NotFound desc = could not find container \"f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f\": container with ID starting with f95f3829d1ed277b3939de0608e01765cd9f4e42651bd2ebafd103478b2dd06f not found: ID does not exist" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.811146 4492 scope.go:117] "RemoveContainer" containerID="1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554" Feb 20 07:38:05 crc kubenswrapper[4492]: E0220 07:38:05.811581 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554\": container with ID starting with 1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554 not found: ID does not exist" containerID="1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.811605 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554"} err="failed to get container status \"1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554\": rpc error: code = NotFound desc = could not find container \"1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554\": container with ID starting with 1872f171160bc6ff1e214d518f6ebe83a712319f79e091315548fe02c7e8c554 not found: ID does not exist" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.811619 4492 scope.go:117] "RemoveContainer" containerID="f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74" Feb 20 07:38:05 crc kubenswrapper[4492]: E0220 07:38:05.811915 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74\": container with ID starting with f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74 not found: ID does not exist" containerID="f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74" Feb 20 07:38:05 crc kubenswrapper[4492]: I0220 07:38:05.811930 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74"} err="failed to get container status \"f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74\": rpc error: code = NotFound desc = could not find container \"f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74\": container with ID starting with f4b7d1d2fa4d98e8d0fd5adb5edfdd524bad34996bac71254afb1334fcd61b74 not found: ID does not exist" Feb 20 07:38:07 crc kubenswrapper[4492]: I0220 07:38:07.565148 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" path="/var/lib/kubelet/pods/3fd6c20f-0afb-48db-b362-724d3ff70682/volumes" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.776666 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jn5t9"] Feb 20 07:38:25 crc kubenswrapper[4492]: E0220 07:38:25.778850 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="extract-utilities" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.778876 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="extract-utilities" Feb 20 07:38:25 crc kubenswrapper[4492]: E0220 07:38:25.778905 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="extract-content" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.778912 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="extract-content" Feb 20 07:38:25 crc kubenswrapper[4492]: E0220 07:38:25.778945 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="registry-server" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.778950 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="registry-server" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.779706 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd6c20f-0afb-48db-b362-724d3ff70682" containerName="registry-server" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.782699 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.800586 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jn5t9"] Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.883767 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2nc8\" (UniqueName: \"kubernetes.io/projected/8e02ea19-1672-4488-ad84-546578dd3132-kube-api-access-r2nc8\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.883855 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-catalog-content\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.883950 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-utilities\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.985019 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2nc8\" (UniqueName: \"kubernetes.io/projected/8e02ea19-1672-4488-ad84-546578dd3132-kube-api-access-r2nc8\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.985110 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-catalog-content\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.985193 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-utilities\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.985510 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-catalog-content\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:25 crc kubenswrapper[4492]: I0220 07:38:25.985795 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-utilities\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:26 crc kubenswrapper[4492]: I0220 07:38:26.012543 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2nc8\" (UniqueName: \"kubernetes.io/projected/8e02ea19-1672-4488-ad84-546578dd3132-kube-api-access-r2nc8\") pod \"community-operators-jn5t9\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:26 crc kubenswrapper[4492]: I0220 07:38:26.100211 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:26 crc kubenswrapper[4492]: I0220 07:38:26.544665 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jn5t9"] Feb 20 07:38:26 crc kubenswrapper[4492]: W0220 07:38:26.556632 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e02ea19_1672_4488_ad84_546578dd3132.slice/crio-48bcb2f1e4c661bf61ee68b3e3abb272787225a5ba392ff4254a358a3832354c WatchSource:0}: Error finding container 48bcb2f1e4c661bf61ee68b3e3abb272787225a5ba392ff4254a358a3832354c: Status 404 returned error can't find the container with id 48bcb2f1e4c661bf61ee68b3e3abb272787225a5ba392ff4254a358a3832354c Feb 20 07:38:26 crc kubenswrapper[4492]: I0220 07:38:26.831383 4492 generic.go:334] "Generic (PLEG): container finished" podID="8e02ea19-1672-4488-ad84-546578dd3132" containerID="c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733" exitCode=0 Feb 20 07:38:26 crc kubenswrapper[4492]: I0220 07:38:26.831591 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jn5t9" event={"ID":"8e02ea19-1672-4488-ad84-546578dd3132","Type":"ContainerDied","Data":"c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733"} Feb 20 07:38:26 crc kubenswrapper[4492]: I0220 07:38:26.831673 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jn5t9" event={"ID":"8e02ea19-1672-4488-ad84-546578dd3132","Type":"ContainerStarted","Data":"48bcb2f1e4c661bf61ee68b3e3abb272787225a5ba392ff4254a358a3832354c"} Feb 20 07:38:27 crc kubenswrapper[4492]: I0220 07:38:27.844519 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jn5t9" event={"ID":"8e02ea19-1672-4488-ad84-546578dd3132","Type":"ContainerStarted","Data":"6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d"} Feb 20 07:38:28 crc kubenswrapper[4492]: I0220 07:38:28.852554 4492 generic.go:334] "Generic (PLEG): container finished" podID="8e02ea19-1672-4488-ad84-546578dd3132" containerID="6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d" exitCode=0 Feb 20 07:38:28 crc kubenswrapper[4492]: I0220 07:38:28.852599 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jn5t9" event={"ID":"8e02ea19-1672-4488-ad84-546578dd3132","Type":"ContainerDied","Data":"6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d"} Feb 20 07:38:29 crc kubenswrapper[4492]: I0220 07:38:29.861322 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jn5t9" event={"ID":"8e02ea19-1672-4488-ad84-546578dd3132","Type":"ContainerStarted","Data":"1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d"} Feb 20 07:38:29 crc kubenswrapper[4492]: I0220 07:38:29.879662 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jn5t9" podStartSLOduration=2.315375189 podStartE2EDuration="4.879259478s" podCreationTimestamp="2026-02-20 07:38:25 +0000 UTC" firstStartedPulling="2026-02-20 07:38:26.835739777 +0000 UTC m=+3463.607028755" lastFinishedPulling="2026-02-20 07:38:29.399624066 +0000 UTC m=+3466.170913044" observedRunningTime="2026-02-20 07:38:29.875562346 +0000 UTC m=+3466.646851334" watchObservedRunningTime="2026-02-20 07:38:29.879259478 +0000 UTC m=+3466.650548456" Feb 20 07:38:36 crc kubenswrapper[4492]: I0220 07:38:36.100807 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:36 crc kubenswrapper[4492]: I0220 07:38:36.101119 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:36 crc kubenswrapper[4492]: I0220 07:38:36.136696 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:36 crc kubenswrapper[4492]: I0220 07:38:36.954813 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:37 crc kubenswrapper[4492]: I0220 07:38:37.001127 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jn5t9"] Feb 20 07:38:38 crc kubenswrapper[4492]: I0220 07:38:38.927626 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jn5t9" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="registry-server" containerID="cri-o://1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d" gracePeriod=2 Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.628890 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.745643 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-utilities\") pod \"8e02ea19-1672-4488-ad84-546578dd3132\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.745699 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2nc8\" (UniqueName: \"kubernetes.io/projected/8e02ea19-1672-4488-ad84-546578dd3132-kube-api-access-r2nc8\") pod \"8e02ea19-1672-4488-ad84-546578dd3132\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.745992 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-catalog-content\") pod \"8e02ea19-1672-4488-ad84-546578dd3132\" (UID: \"8e02ea19-1672-4488-ad84-546578dd3132\") " Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.746241 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-utilities" (OuterVolumeSpecName: "utilities") pod "8e02ea19-1672-4488-ad84-546578dd3132" (UID: "8e02ea19-1672-4488-ad84-546578dd3132"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.755282 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e02ea19-1672-4488-ad84-546578dd3132-kube-api-access-r2nc8" (OuterVolumeSpecName: "kube-api-access-r2nc8") pod "8e02ea19-1672-4488-ad84-546578dd3132" (UID: "8e02ea19-1672-4488-ad84-546578dd3132"). InnerVolumeSpecName "kube-api-access-r2nc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.828675 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e02ea19-1672-4488-ad84-546578dd3132" (UID: "8e02ea19-1672-4488-ad84-546578dd3132"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.848036 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.848060 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e02ea19-1672-4488-ad84-546578dd3132-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.848068 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2nc8\" (UniqueName: \"kubernetes.io/projected/8e02ea19-1672-4488-ad84-546578dd3132-kube-api-access-r2nc8\") on node \"crc\" DevicePath \"\"" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.934458 4492 generic.go:334] "Generic (PLEG): container finished" podID="8e02ea19-1672-4488-ad84-546578dd3132" containerID="1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d" exitCode=0 Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.934529 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jn5t9" event={"ID":"8e02ea19-1672-4488-ad84-546578dd3132","Type":"ContainerDied","Data":"1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d"} Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.934580 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jn5t9" event={"ID":"8e02ea19-1672-4488-ad84-546578dd3132","Type":"ContainerDied","Data":"48bcb2f1e4c661bf61ee68b3e3abb272787225a5ba392ff4254a358a3832354c"} Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.934600 4492 scope.go:117] "RemoveContainer" containerID="1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.935034 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jn5t9" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.965569 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jn5t9"] Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.966286 4492 scope.go:117] "RemoveContainer" containerID="6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d" Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.974856 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jn5t9"] Feb 20 07:38:39 crc kubenswrapper[4492]: I0220 07:38:39.984873 4492 scope.go:117] "RemoveContainer" containerID="c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733" Feb 20 07:38:40 crc kubenswrapper[4492]: I0220 07:38:40.022180 4492 scope.go:117] "RemoveContainer" containerID="1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d" Feb 20 07:38:40 crc kubenswrapper[4492]: E0220 07:38:40.023582 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d\": container with ID starting with 1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d not found: ID does not exist" containerID="1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d" Feb 20 07:38:40 crc kubenswrapper[4492]: I0220 07:38:40.024171 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d"} err="failed to get container status \"1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d\": rpc error: code = NotFound desc = could not find container \"1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d\": container with ID starting with 1e7aa1e73277dad738b288fbc7ab9687ebc0057a89d0941986b5d67b611fe56d not found: ID does not exist" Feb 20 07:38:40 crc kubenswrapper[4492]: I0220 07:38:40.024207 4492 scope.go:117] "RemoveContainer" containerID="6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d" Feb 20 07:38:40 crc kubenswrapper[4492]: E0220 07:38:40.024511 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d\": container with ID starting with 6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d not found: ID does not exist" containerID="6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d" Feb 20 07:38:40 crc kubenswrapper[4492]: I0220 07:38:40.024534 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d"} err="failed to get container status \"6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d\": rpc error: code = NotFound desc = could not find container \"6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d\": container with ID starting with 6f5fb45e2f43a9d292ca324a72db7c8951c1a5133cfb1d477b5a69245b8e467d not found: ID does not exist" Feb 20 07:38:40 crc kubenswrapper[4492]: I0220 07:38:40.024548 4492 scope.go:117] "RemoveContainer" containerID="c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733" Feb 20 07:38:40 crc kubenswrapper[4492]: E0220 07:38:40.024833 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733\": container with ID starting with c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733 not found: ID does not exist" containerID="c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733" Feb 20 07:38:40 crc kubenswrapper[4492]: I0220 07:38:40.024860 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733"} err="failed to get container status \"c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733\": rpc error: code = NotFound desc = could not find container \"c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733\": container with ID starting with c421e47b7f73c78329e2aa1dfa0deefb23c23c96349dc74f876f2f8aa6a84733 not found: ID does not exist" Feb 20 07:38:41 crc kubenswrapper[4492]: I0220 07:38:41.564292 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e02ea19-1672-4488-ad84-546578dd3132" path="/var/lib/kubelet/pods/8e02ea19-1672-4488-ad84-546578dd3132/volumes" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.723221 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5pgjq"] Feb 20 07:39:07 crc kubenswrapper[4492]: E0220 07:39:07.725987 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="registry-server" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.726017 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="registry-server" Feb 20 07:39:07 crc kubenswrapper[4492]: E0220 07:39:07.726048 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="extract-content" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.726054 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="extract-content" Feb 20 07:39:07 crc kubenswrapper[4492]: E0220 07:39:07.726066 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="extract-utilities" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.726071 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="extract-utilities" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.726279 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e02ea19-1672-4488-ad84-546578dd3132" containerName="registry-server" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.728341 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.736007 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pgjq"] Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.801870 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-utilities\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.802037 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-catalog-content\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.802135 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmf9g\" (UniqueName: \"kubernetes.io/projected/4455b521-86e2-4876-9021-4def2842ea86-kube-api-access-qmf9g\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.903984 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-utilities\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.904029 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-catalog-content\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.904060 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmf9g\" (UniqueName: \"kubernetes.io/projected/4455b521-86e2-4876-9021-4def2842ea86-kube-api-access-qmf9g\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.905289 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-utilities\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.905658 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-catalog-content\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:07 crc kubenswrapper[4492]: I0220 07:39:07.924918 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmf9g\" (UniqueName: \"kubernetes.io/projected/4455b521-86e2-4876-9021-4def2842ea86-kube-api-access-qmf9g\") pod \"redhat-operators-5pgjq\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:08 crc kubenswrapper[4492]: I0220 07:39:08.048787 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:08 crc kubenswrapper[4492]: I0220 07:39:08.458357 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pgjq"] Feb 20 07:39:09 crc kubenswrapper[4492]: I0220 07:39:09.123915 4492 generic.go:334] "Generic (PLEG): container finished" podID="4455b521-86e2-4876-9021-4def2842ea86" containerID="60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5" exitCode=0 Feb 20 07:39:09 crc kubenswrapper[4492]: I0220 07:39:09.123969 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pgjq" event={"ID":"4455b521-86e2-4876-9021-4def2842ea86","Type":"ContainerDied","Data":"60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5"} Feb 20 07:39:09 crc kubenswrapper[4492]: I0220 07:39:09.124190 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pgjq" event={"ID":"4455b521-86e2-4876-9021-4def2842ea86","Type":"ContainerStarted","Data":"b01f754961ba75916e138e937c4df01cd3f9a6fef4cac846b698948a923234ac"} Feb 20 07:39:10 crc kubenswrapper[4492]: I0220 07:39:10.132511 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pgjq" event={"ID":"4455b521-86e2-4876-9021-4def2842ea86","Type":"ContainerStarted","Data":"8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983"} Feb 20 07:39:14 crc kubenswrapper[4492]: I0220 07:39:14.158728 4492 generic.go:334] "Generic (PLEG): container finished" podID="4455b521-86e2-4876-9021-4def2842ea86" containerID="8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983" exitCode=0 Feb 20 07:39:14 crc kubenswrapper[4492]: I0220 07:39:14.158820 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pgjq" event={"ID":"4455b521-86e2-4876-9021-4def2842ea86","Type":"ContainerDied","Data":"8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983"} Feb 20 07:39:15 crc kubenswrapper[4492]: I0220 07:39:15.168022 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pgjq" event={"ID":"4455b521-86e2-4876-9021-4def2842ea86","Type":"ContainerStarted","Data":"99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1"} Feb 20 07:39:15 crc kubenswrapper[4492]: I0220 07:39:15.188196 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5pgjq" podStartSLOduration=2.646378631 podStartE2EDuration="8.188167572s" podCreationTimestamp="2026-02-20 07:39:07 +0000 UTC" firstStartedPulling="2026-02-20 07:39:09.125604839 +0000 UTC m=+3505.896893818" lastFinishedPulling="2026-02-20 07:39:14.667393781 +0000 UTC m=+3511.438682759" observedRunningTime="2026-02-20 07:39:15.18158558 +0000 UTC m=+3511.952874557" watchObservedRunningTime="2026-02-20 07:39:15.188167572 +0000 UTC m=+3511.959456550" Feb 20 07:39:18 crc kubenswrapper[4492]: I0220 07:39:18.049809 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:18 crc kubenswrapper[4492]: I0220 07:39:18.050198 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:19 crc kubenswrapper[4492]: I0220 07:39:19.084407 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5pgjq" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="registry-server" probeResult="failure" output=< Feb 20 07:39:19 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:39:19 crc kubenswrapper[4492]: > Feb 20 07:39:29 crc kubenswrapper[4492]: I0220 07:39:29.079374 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5pgjq" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="registry-server" probeResult="failure" output=< Feb 20 07:39:29 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:39:29 crc kubenswrapper[4492]: > Feb 20 07:39:38 crc kubenswrapper[4492]: I0220 07:39:38.083470 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:38 crc kubenswrapper[4492]: I0220 07:39:38.118572 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:38 crc kubenswrapper[4492]: I0220 07:39:38.920136 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pgjq"] Feb 20 07:39:39 crc kubenswrapper[4492]: I0220 07:39:39.311710 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:39:39 crc kubenswrapper[4492]: I0220 07:39:39.312905 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:39:39 crc kubenswrapper[4492]: I0220 07:39:39.323546 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5pgjq" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="registry-server" containerID="cri-o://99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1" gracePeriod=2 Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.123882 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.244968 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmf9g\" (UniqueName: \"kubernetes.io/projected/4455b521-86e2-4876-9021-4def2842ea86-kube-api-access-qmf9g\") pod \"4455b521-86e2-4876-9021-4def2842ea86\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.245080 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-catalog-content\") pod \"4455b521-86e2-4876-9021-4def2842ea86\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.245200 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-utilities\") pod \"4455b521-86e2-4876-9021-4def2842ea86\" (UID: \"4455b521-86e2-4876-9021-4def2842ea86\") " Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.247534 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-utilities" (OuterVolumeSpecName: "utilities") pod "4455b521-86e2-4876-9021-4def2842ea86" (UID: "4455b521-86e2-4876-9021-4def2842ea86"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.262434 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4455b521-86e2-4876-9021-4def2842ea86-kube-api-access-qmf9g" (OuterVolumeSpecName: "kube-api-access-qmf9g") pod "4455b521-86e2-4876-9021-4def2842ea86" (UID: "4455b521-86e2-4876-9021-4def2842ea86"). InnerVolumeSpecName "kube-api-access-qmf9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.330431 4492 generic.go:334] "Generic (PLEG): container finished" podID="4455b521-86e2-4876-9021-4def2842ea86" containerID="99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1" exitCode=0 Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.330495 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pgjq" event={"ID":"4455b521-86e2-4876-9021-4def2842ea86","Type":"ContainerDied","Data":"99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1"} Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.330525 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pgjq" event={"ID":"4455b521-86e2-4876-9021-4def2842ea86","Type":"ContainerDied","Data":"b01f754961ba75916e138e937c4df01cd3f9a6fef4cac846b698948a923234ac"} Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.330694 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pgjq" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.331670 4492 scope.go:117] "RemoveContainer" containerID="99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.348442 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmf9g\" (UniqueName: \"kubernetes.io/projected/4455b521-86e2-4876-9021-4def2842ea86-kube-api-access-qmf9g\") on node \"crc\" DevicePath \"\"" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.348493 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.363100 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4455b521-86e2-4876-9021-4def2842ea86" (UID: "4455b521-86e2-4876-9021-4def2842ea86"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.370582 4492 scope.go:117] "RemoveContainer" containerID="8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.386785 4492 scope.go:117] "RemoveContainer" containerID="60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.422396 4492 scope.go:117] "RemoveContainer" containerID="99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1" Feb 20 07:39:40 crc kubenswrapper[4492]: E0220 07:39:40.426668 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1\": container with ID starting with 99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1 not found: ID does not exist" containerID="99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.427263 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1"} err="failed to get container status \"99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1\": rpc error: code = NotFound desc = could not find container \"99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1\": container with ID starting with 99e19cd42a42383934a25fb10569e80885ff23192ca2fcb58dc5ce081ea77ae1 not found: ID does not exist" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.427299 4492 scope.go:117] "RemoveContainer" containerID="8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983" Feb 20 07:39:40 crc kubenswrapper[4492]: E0220 07:39:40.427807 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983\": container with ID starting with 8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983 not found: ID does not exist" containerID="8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.427859 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983"} err="failed to get container status \"8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983\": rpc error: code = NotFound desc = could not find container \"8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983\": container with ID starting with 8ee9734220d491f75dcaba1c59e4e012dfa949d3d22c5f01292284b753395983 not found: ID does not exist" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.427884 4492 scope.go:117] "RemoveContainer" containerID="60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5" Feb 20 07:39:40 crc kubenswrapper[4492]: E0220 07:39:40.428210 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5\": container with ID starting with 60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5 not found: ID does not exist" containerID="60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.428232 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5"} err="failed to get container status \"60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5\": rpc error: code = NotFound desc = could not find container \"60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5\": container with ID starting with 60a6921a97d38e0161ed5a656f3f10fd6d3b4850c406ef67538322a9fd2a3ec5 not found: ID does not exist" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.450445 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4455b521-86e2-4876-9021-4def2842ea86-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.659091 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pgjq"] Feb 20 07:39:40 crc kubenswrapper[4492]: I0220 07:39:40.665722 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5pgjq"] Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.544469 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vmcxc"] Feb 20 07:39:41 crc kubenswrapper[4492]: E0220 07:39:41.546167 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="registry-server" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.546193 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="registry-server" Feb 20 07:39:41 crc kubenswrapper[4492]: E0220 07:39:41.546217 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="extract-content" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.546223 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="extract-content" Feb 20 07:39:41 crc kubenswrapper[4492]: E0220 07:39:41.546242 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="extract-utilities" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.546247 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="extract-utilities" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.547006 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="4455b521-86e2-4876-9021-4def2842ea86" containerName="registry-server" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.551323 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.565928 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4455b521-86e2-4876-9021-4def2842ea86" path="/var/lib/kubelet/pods/4455b521-86e2-4876-9021-4def2842ea86/volumes" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.573316 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmcxc"] Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.670974 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-utilities\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.671210 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c8nt\" (UniqueName: \"kubernetes.io/projected/dae44db6-0deb-493e-889b-aefbb855fb6e-kube-api-access-2c8nt\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.671380 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-catalog-content\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.773124 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c8nt\" (UniqueName: \"kubernetes.io/projected/dae44db6-0deb-493e-889b-aefbb855fb6e-kube-api-access-2c8nt\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.773235 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-catalog-content\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.773262 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-utilities\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.773629 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-catalog-content\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.773711 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-utilities\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.787405 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c8nt\" (UniqueName: \"kubernetes.io/projected/dae44db6-0deb-493e-889b-aefbb855fb6e-kube-api-access-2c8nt\") pod \"redhat-marketplace-vmcxc\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:41 crc kubenswrapper[4492]: I0220 07:39:41.871896 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:42 crc kubenswrapper[4492]: I0220 07:39:42.410315 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmcxc"] Feb 20 07:39:43 crc kubenswrapper[4492]: I0220 07:39:43.352553 4492 generic.go:334] "Generic (PLEG): container finished" podID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerID="4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c" exitCode=0 Feb 20 07:39:43 crc kubenswrapper[4492]: I0220 07:39:43.352639 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmcxc" event={"ID":"dae44db6-0deb-493e-889b-aefbb855fb6e","Type":"ContainerDied","Data":"4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c"} Feb 20 07:39:43 crc kubenswrapper[4492]: I0220 07:39:43.353510 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmcxc" event={"ID":"dae44db6-0deb-493e-889b-aefbb855fb6e","Type":"ContainerStarted","Data":"c389e94c6d8c7c3a8ee9e624692fdd3e81c0a9beea782ae41898f0b046a1f9e4"} Feb 20 07:39:44 crc kubenswrapper[4492]: I0220 07:39:44.362016 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmcxc" event={"ID":"dae44db6-0deb-493e-889b-aefbb855fb6e","Type":"ContainerStarted","Data":"bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7"} Feb 20 07:39:45 crc kubenswrapper[4492]: I0220 07:39:45.373261 4492 generic.go:334] "Generic (PLEG): container finished" podID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerID="bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7" exitCode=0 Feb 20 07:39:45 crc kubenswrapper[4492]: I0220 07:39:45.373468 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmcxc" event={"ID":"dae44db6-0deb-493e-889b-aefbb855fb6e","Type":"ContainerDied","Data":"bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7"} Feb 20 07:39:46 crc kubenswrapper[4492]: I0220 07:39:46.383701 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmcxc" event={"ID":"dae44db6-0deb-493e-889b-aefbb855fb6e","Type":"ContainerStarted","Data":"f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1"} Feb 20 07:39:46 crc kubenswrapper[4492]: I0220 07:39:46.398901 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vmcxc" podStartSLOduration=2.894212044 podStartE2EDuration="5.398888698s" podCreationTimestamp="2026-02-20 07:39:41 +0000 UTC" firstStartedPulling="2026-02-20 07:39:43.354084967 +0000 UTC m=+3540.125373944" lastFinishedPulling="2026-02-20 07:39:45.85876162 +0000 UTC m=+3542.630050598" observedRunningTime="2026-02-20 07:39:46.398845877 +0000 UTC m=+3543.170134856" watchObservedRunningTime="2026-02-20 07:39:46.398888698 +0000 UTC m=+3543.170177676" Feb 20 07:39:51 crc kubenswrapper[4492]: I0220 07:39:51.872980 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:51 crc kubenswrapper[4492]: I0220 07:39:51.873373 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:51 crc kubenswrapper[4492]: I0220 07:39:51.906519 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:52 crc kubenswrapper[4492]: I0220 07:39:52.453853 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:52 crc kubenswrapper[4492]: I0220 07:39:52.491740 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmcxc"] Feb 20 07:39:54 crc kubenswrapper[4492]: I0220 07:39:54.436325 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vmcxc" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="registry-server" containerID="cri-o://f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1" gracePeriod=2 Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.002322 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.111283 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-utilities\") pod \"dae44db6-0deb-493e-889b-aefbb855fb6e\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.111585 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c8nt\" (UniqueName: \"kubernetes.io/projected/dae44db6-0deb-493e-889b-aefbb855fb6e-kube-api-access-2c8nt\") pod \"dae44db6-0deb-493e-889b-aefbb855fb6e\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.111792 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-catalog-content\") pod \"dae44db6-0deb-493e-889b-aefbb855fb6e\" (UID: \"dae44db6-0deb-493e-889b-aefbb855fb6e\") " Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.111864 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-utilities" (OuterVolumeSpecName: "utilities") pod "dae44db6-0deb-493e-889b-aefbb855fb6e" (UID: "dae44db6-0deb-493e-889b-aefbb855fb6e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.113673 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.119593 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dae44db6-0deb-493e-889b-aefbb855fb6e-kube-api-access-2c8nt" (OuterVolumeSpecName: "kube-api-access-2c8nt") pod "dae44db6-0deb-493e-889b-aefbb855fb6e" (UID: "dae44db6-0deb-493e-889b-aefbb855fb6e"). InnerVolumeSpecName "kube-api-access-2c8nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.128593 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dae44db6-0deb-493e-889b-aefbb855fb6e" (UID: "dae44db6-0deb-493e-889b-aefbb855fb6e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.216072 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dae44db6-0deb-493e-889b-aefbb855fb6e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.216103 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c8nt\" (UniqueName: \"kubernetes.io/projected/dae44db6-0deb-493e-889b-aefbb855fb6e-kube-api-access-2c8nt\") on node \"crc\" DevicePath \"\"" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.450177 4492 generic.go:334] "Generic (PLEG): container finished" podID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerID="f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1" exitCode=0 Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.450218 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmcxc" event={"ID":"dae44db6-0deb-493e-889b-aefbb855fb6e","Type":"ContainerDied","Data":"f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1"} Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.450246 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vmcxc" event={"ID":"dae44db6-0deb-493e-889b-aefbb855fb6e","Type":"ContainerDied","Data":"c389e94c6d8c7c3a8ee9e624692fdd3e81c0a9beea782ae41898f0b046a1f9e4"} Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.450265 4492 scope.go:117] "RemoveContainer" containerID="f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.450397 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vmcxc" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.488919 4492 scope.go:117] "RemoveContainer" containerID="bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.491164 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmcxc"] Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.497128 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vmcxc"] Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.505299 4492 scope.go:117] "RemoveContainer" containerID="4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.539199 4492 scope.go:117] "RemoveContainer" containerID="f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1" Feb 20 07:39:55 crc kubenswrapper[4492]: E0220 07:39:55.539523 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1\": container with ID starting with f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1 not found: ID does not exist" containerID="f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.539562 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1"} err="failed to get container status \"f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1\": rpc error: code = NotFound desc = could not find container \"f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1\": container with ID starting with f1062cd7d6f49d2669897a18ad44f3fa88c5024b365e5473bd296052eccdeec1 not found: ID does not exist" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.539589 4492 scope.go:117] "RemoveContainer" containerID="bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7" Feb 20 07:39:55 crc kubenswrapper[4492]: E0220 07:39:55.541051 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7\": container with ID starting with bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7 not found: ID does not exist" containerID="bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.541080 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7"} err="failed to get container status \"bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7\": rpc error: code = NotFound desc = could not find container \"bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7\": container with ID starting with bdd4b80b23b6d4e42c66d73e83bb1ba04bf1940a1fadd929889b29ae72352dd7 not found: ID does not exist" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.541098 4492 scope.go:117] "RemoveContainer" containerID="4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c" Feb 20 07:39:55 crc kubenswrapper[4492]: E0220 07:39:55.541332 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c\": container with ID starting with 4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c not found: ID does not exist" containerID="4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.541350 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c"} err="failed to get container status \"4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c\": rpc error: code = NotFound desc = could not find container \"4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c\": container with ID starting with 4dfa20bab8f77dfb9133ab4ed86d0786787992f5c43be7557e8e23d09feb052c not found: ID does not exist" Feb 20 07:39:55 crc kubenswrapper[4492]: I0220 07:39:55.569933 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" path="/var/lib/kubelet/pods/dae44db6-0deb-493e-889b-aefbb855fb6e/volumes" Feb 20 07:40:09 crc kubenswrapper[4492]: I0220 07:40:09.311338 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:40:09 crc kubenswrapper[4492]: I0220 07:40:09.312025 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.310892 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.311730 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.311805 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.313134 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.313232 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" gracePeriod=600 Feb 20 07:40:39 crc kubenswrapper[4492]: E0220 07:40:39.434508 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.855264 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" exitCode=0 Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.855307 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165"} Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.855438 4492 scope.go:117] "RemoveContainer" containerID="172222d62e8e3557fb3aeaadcdb3793eab777225daeaf180018aac17f6dcf60d" Feb 20 07:40:39 crc kubenswrapper[4492]: I0220 07:40:39.856883 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:40:39 crc kubenswrapper[4492]: E0220 07:40:39.857623 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:40:51 crc kubenswrapper[4492]: I0220 07:40:51.558993 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:40:51 crc kubenswrapper[4492]: E0220 07:40:51.559871 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:41:03 crc kubenswrapper[4492]: I0220 07:41:03.570874 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:41:03 crc kubenswrapper[4492]: E0220 07:41:03.574267 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:41:16 crc kubenswrapper[4492]: I0220 07:41:16.557730 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:41:16 crc kubenswrapper[4492]: E0220 07:41:16.558567 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:41:30 crc kubenswrapper[4492]: I0220 07:41:30.557162 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:41:30 crc kubenswrapper[4492]: E0220 07:41:30.558100 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:41:44 crc kubenswrapper[4492]: I0220 07:41:44.557725 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:41:44 crc kubenswrapper[4492]: E0220 07:41:44.558805 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:41:57 crc kubenswrapper[4492]: I0220 07:41:57.557009 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:41:57 crc kubenswrapper[4492]: E0220 07:41:57.557579 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:42:10 crc kubenswrapper[4492]: I0220 07:42:10.557435 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:42:10 crc kubenswrapper[4492]: E0220 07:42:10.559686 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:42:25 crc kubenswrapper[4492]: I0220 07:42:25.558064 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:42:25 crc kubenswrapper[4492]: E0220 07:42:25.558921 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:42:40 crc kubenswrapper[4492]: I0220 07:42:40.556455 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:42:40 crc kubenswrapper[4492]: E0220 07:42:40.557306 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:42:52 crc kubenswrapper[4492]: I0220 07:42:52.557960 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:42:52 crc kubenswrapper[4492]: E0220 07:42:52.558768 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:43:07 crc kubenswrapper[4492]: I0220 07:43:07.557533 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:43:07 crc kubenswrapper[4492]: E0220 07:43:07.558555 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:43:18 crc kubenswrapper[4492]: I0220 07:43:18.556389 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:43:18 crc kubenswrapper[4492]: E0220 07:43:18.557309 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:43:32 crc kubenswrapper[4492]: I0220 07:43:32.557895 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:43:32 crc kubenswrapper[4492]: E0220 07:43:32.559007 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:43:44 crc kubenswrapper[4492]: I0220 07:43:44.558370 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:43:44 crc kubenswrapper[4492]: E0220 07:43:44.560972 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:43:58 crc kubenswrapper[4492]: I0220 07:43:58.558778 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:43:58 crc kubenswrapper[4492]: E0220 07:43:58.559598 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:44:11 crc kubenswrapper[4492]: I0220 07:44:11.559008 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:44:11 crc kubenswrapper[4492]: E0220 07:44:11.559883 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:44:22 crc kubenswrapper[4492]: I0220 07:44:22.557432 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:44:22 crc kubenswrapper[4492]: E0220 07:44:22.558031 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:44:35 crc kubenswrapper[4492]: I0220 07:44:35.557301 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:44:35 crc kubenswrapper[4492]: E0220 07:44:35.557920 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:44:46 crc kubenswrapper[4492]: I0220 07:44:46.556307 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:44:46 crc kubenswrapper[4492]: E0220 07:44:46.558007 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:44:59 crc kubenswrapper[4492]: I0220 07:44:59.557777 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:44:59 crc kubenswrapper[4492]: E0220 07:44:59.558854 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.231984 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp"] Feb 20 07:45:00 crc kubenswrapper[4492]: E0220 07:45:00.234328 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="extract-content" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.234362 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="extract-content" Feb 20 07:45:00 crc kubenswrapper[4492]: E0220 07:45:00.234398 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="extract-utilities" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.234405 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="extract-utilities" Feb 20 07:45:00 crc kubenswrapper[4492]: E0220 07:45:00.234434 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="registry-server" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.234442 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="registry-server" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.234761 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="dae44db6-0deb-493e-889b-aefbb855fb6e" containerName="registry-server" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.235541 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.245326 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.245331 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.266171 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp"] Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.342966 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc2zb\" (UniqueName: \"kubernetes.io/projected/80bab9bc-ffef-40d6-8a61-a834994983d5-kube-api-access-fc2zb\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.343550 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80bab9bc-ffef-40d6-8a61-a834994983d5-secret-volume\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.343732 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80bab9bc-ffef-40d6-8a61-a834994983d5-config-volume\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.447571 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc2zb\" (UniqueName: \"kubernetes.io/projected/80bab9bc-ffef-40d6-8a61-a834994983d5-kube-api-access-fc2zb\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.448070 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80bab9bc-ffef-40d6-8a61-a834994983d5-secret-volume\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.448538 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80bab9bc-ffef-40d6-8a61-a834994983d5-config-volume\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.449334 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80bab9bc-ffef-40d6-8a61-a834994983d5-config-volume\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.456862 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80bab9bc-ffef-40d6-8a61-a834994983d5-secret-volume\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.461929 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc2zb\" (UniqueName: \"kubernetes.io/projected/80bab9bc-ffef-40d6-8a61-a834994983d5-kube-api-access-fc2zb\") pod \"collect-profiles-29526225-5zqhp\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.556620 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:00 crc kubenswrapper[4492]: I0220 07:45:00.991650 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp"] Feb 20 07:45:01 crc kubenswrapper[4492]: I0220 07:45:01.973245 4492 generic.go:334] "Generic (PLEG): container finished" podID="80bab9bc-ffef-40d6-8a61-a834994983d5" containerID="a129bb0d8c8469d6031580ca1e3f7b1ad9a4ae0e0f35bb28288082d423ead4fd" exitCode=0 Feb 20 07:45:01 crc kubenswrapper[4492]: I0220 07:45:01.973326 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" event={"ID":"80bab9bc-ffef-40d6-8a61-a834994983d5","Type":"ContainerDied","Data":"a129bb0d8c8469d6031580ca1e3f7b1ad9a4ae0e0f35bb28288082d423ead4fd"} Feb 20 07:45:01 crc kubenswrapper[4492]: I0220 07:45:01.973637 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" event={"ID":"80bab9bc-ffef-40d6-8a61-a834994983d5","Type":"ContainerStarted","Data":"cf5c66ad8817807ba4ea6dd3aa228024bb67b3a69069c86a9799741d2f9ef7eb"} Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.401661 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.516311 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80bab9bc-ffef-40d6-8a61-a834994983d5-config-volume\") pod \"80bab9bc-ffef-40d6-8a61-a834994983d5\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.516533 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80bab9bc-ffef-40d6-8a61-a834994983d5-secret-volume\") pod \"80bab9bc-ffef-40d6-8a61-a834994983d5\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.516663 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc2zb\" (UniqueName: \"kubernetes.io/projected/80bab9bc-ffef-40d6-8a61-a834994983d5-kube-api-access-fc2zb\") pod \"80bab9bc-ffef-40d6-8a61-a834994983d5\" (UID: \"80bab9bc-ffef-40d6-8a61-a834994983d5\") " Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.517759 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80bab9bc-ffef-40d6-8a61-a834994983d5-config-volume" (OuterVolumeSpecName: "config-volume") pod "80bab9bc-ffef-40d6-8a61-a834994983d5" (UID: "80bab9bc-ffef-40d6-8a61-a834994983d5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.523025 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80bab9bc-ffef-40d6-8a61-a834994983d5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "80bab9bc-ffef-40d6-8a61-a834994983d5" (UID: "80bab9bc-ffef-40d6-8a61-a834994983d5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.523116 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80bab9bc-ffef-40d6-8a61-a834994983d5-kube-api-access-fc2zb" (OuterVolumeSpecName: "kube-api-access-fc2zb") pod "80bab9bc-ffef-40d6-8a61-a834994983d5" (UID: "80bab9bc-ffef-40d6-8a61-a834994983d5"). InnerVolumeSpecName "kube-api-access-fc2zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.619546 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80bab9bc-ffef-40d6-8a61-a834994983d5-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.619578 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80bab9bc-ffef-40d6-8a61-a834994983d5-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.619589 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc2zb\" (UniqueName: \"kubernetes.io/projected/80bab9bc-ffef-40d6-8a61-a834994983d5-kube-api-access-fc2zb\") on node \"crc\" DevicePath \"\"" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.986998 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" event={"ID":"80bab9bc-ffef-40d6-8a61-a834994983d5","Type":"ContainerDied","Data":"cf5c66ad8817807ba4ea6dd3aa228024bb67b3a69069c86a9799741d2f9ef7eb"} Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.987061 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp" Feb 20 07:45:03 crc kubenswrapper[4492]: I0220 07:45:03.987043 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf5c66ad8817807ba4ea6dd3aa228024bb67b3a69069c86a9799741d2f9ef7eb" Feb 20 07:45:04 crc kubenswrapper[4492]: I0220 07:45:04.470291 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms"] Feb 20 07:45:04 crc kubenswrapper[4492]: I0220 07:45:04.478420 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526180-lqsms"] Feb 20 07:45:05 crc kubenswrapper[4492]: I0220 07:45:05.565338 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ad0831-93b1-4f58-9a4b-772d16f072c2" path="/var/lib/kubelet/pods/09ad0831-93b1-4f58-9a4b-772d16f072c2/volumes" Feb 20 07:45:07 crc kubenswrapper[4492]: I0220 07:45:07.972434 4492 scope.go:117] "RemoveContainer" containerID="a82787a5e86af1abdaa5700884c2a68da34c7a9702d857e4cc1571e00a22be5b" Feb 20 07:45:11 crc kubenswrapper[4492]: I0220 07:45:11.557180 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:45:11 crc kubenswrapper[4492]: E0220 07:45:11.557949 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:45:26 crc kubenswrapper[4492]: I0220 07:45:26.557539 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:45:26 crc kubenswrapper[4492]: E0220 07:45:26.558708 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:45:39 crc kubenswrapper[4492]: I0220 07:45:39.557695 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:45:40 crc kubenswrapper[4492]: I0220 07:45:40.299829 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"4df0d1375018eccda77abe46780f42ef0dc35696ecbd3e7a7ca7cffebddee9ed"} Feb 20 07:47:39 crc kubenswrapper[4492]: I0220 07:47:39.311312 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:47:39 crc kubenswrapper[4492]: I0220 07:47:39.312614 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:48:09 crc kubenswrapper[4492]: I0220 07:48:09.311386 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:48:09 crc kubenswrapper[4492]: I0220 07:48:09.312571 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.128949 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-67rg8"] Feb 20 07:48:29 crc kubenswrapper[4492]: E0220 07:48:29.134268 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80bab9bc-ffef-40d6-8a61-a834994983d5" containerName="collect-profiles" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.134298 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="80bab9bc-ffef-40d6-8a61-a834994983d5" containerName="collect-profiles" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.142499 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="80bab9bc-ffef-40d6-8a61-a834994983d5" containerName="collect-profiles" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.183731 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-67rg8"] Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.185530 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.211349 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922456bc-c937-4182-b66c-9b6cc71f80a4-utilities\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.211558 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922456bc-c937-4182-b66c-9b6cc71f80a4-catalog-content\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.211701 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm4m9\" (UniqueName: \"kubernetes.io/projected/922456bc-c937-4182-b66c-9b6cc71f80a4-kube-api-access-dm4m9\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.312946 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922456bc-c937-4182-b66c-9b6cc71f80a4-catalog-content\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.313072 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm4m9\" (UniqueName: \"kubernetes.io/projected/922456bc-c937-4182-b66c-9b6cc71f80a4-kube-api-access-dm4m9\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.313216 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922456bc-c937-4182-b66c-9b6cc71f80a4-utilities\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.316581 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922456bc-c937-4182-b66c-9b6cc71f80a4-utilities\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.316582 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922456bc-c937-4182-b66c-9b6cc71f80a4-catalog-content\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.340024 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm4m9\" (UniqueName: \"kubernetes.io/projected/922456bc-c937-4182-b66c-9b6cc71f80a4-kube-api-access-dm4m9\") pod \"community-operators-67rg8\" (UID: \"922456bc-c937-4182-b66c-9b6cc71f80a4\") " pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:29 crc kubenswrapper[4492]: I0220 07:48:29.533127 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:30 crc kubenswrapper[4492]: I0220 07:48:30.206337 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-67rg8"] Feb 20 07:48:30 crc kubenswrapper[4492]: I0220 07:48:30.705222 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67rg8" event={"ID":"922456bc-c937-4182-b66c-9b6cc71f80a4","Type":"ContainerDied","Data":"b79b2810c0b099b82e7745ee299a7fb83821deeb2fd55ef3581d2d5c1c92740d"} Feb 20 07:48:30 crc kubenswrapper[4492]: I0220 07:48:30.706081 4492 generic.go:334] "Generic (PLEG): container finished" podID="922456bc-c937-4182-b66c-9b6cc71f80a4" containerID="b79b2810c0b099b82e7745ee299a7fb83821deeb2fd55ef3581d2d5c1c92740d" exitCode=0 Feb 20 07:48:30 crc kubenswrapper[4492]: I0220 07:48:30.706551 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67rg8" event={"ID":"922456bc-c937-4182-b66c-9b6cc71f80a4","Type":"ContainerStarted","Data":"5ee4232720d2af47561001e3f27edd5922d164c152f9e4f25ee1ad58b0bae5af"} Feb 20 07:48:30 crc kubenswrapper[4492]: I0220 07:48:30.713724 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:48:35 crc kubenswrapper[4492]: I0220 07:48:35.761860 4492 generic.go:334] "Generic (PLEG): container finished" podID="922456bc-c937-4182-b66c-9b6cc71f80a4" containerID="ac562bb4c65f1212bef5aeb5bbd0abfae386c92675332cb671cacf45077a1c09" exitCode=0 Feb 20 07:48:35 crc kubenswrapper[4492]: I0220 07:48:35.762012 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67rg8" event={"ID":"922456bc-c937-4182-b66c-9b6cc71f80a4","Type":"ContainerDied","Data":"ac562bb4c65f1212bef5aeb5bbd0abfae386c92675332cb671cacf45077a1c09"} Feb 20 07:48:36 crc kubenswrapper[4492]: I0220 07:48:36.773333 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-67rg8" event={"ID":"922456bc-c937-4182-b66c-9b6cc71f80a4","Type":"ContainerStarted","Data":"bf29fb601c074779f451584422416ea670c96ff58b404f09c30a5456780d0dbc"} Feb 20 07:48:36 crc kubenswrapper[4492]: I0220 07:48:36.799868 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-67rg8" podStartSLOduration=2.207205628 podStartE2EDuration="7.798021136s" podCreationTimestamp="2026-02-20 07:48:29 +0000 UTC" firstStartedPulling="2026-02-20 07:48:30.708394245 +0000 UTC m=+4067.479683223" lastFinishedPulling="2026-02-20 07:48:36.299209754 +0000 UTC m=+4073.070498731" observedRunningTime="2026-02-20 07:48:36.792005582 +0000 UTC m=+4073.563294559" watchObservedRunningTime="2026-02-20 07:48:36.798021136 +0000 UTC m=+4073.569310114" Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.312036 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.313390 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.313459 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.314367 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4df0d1375018eccda77abe46780f42ef0dc35696ecbd3e7a7ca7cffebddee9ed"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.314417 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://4df0d1375018eccda77abe46780f42ef0dc35696ecbd3e7a7ca7cffebddee9ed" gracePeriod=600 Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.533946 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.534234 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.798592 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="4df0d1375018eccda77abe46780f42ef0dc35696ecbd3e7a7ca7cffebddee9ed" exitCode=0 Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.798668 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"4df0d1375018eccda77abe46780f42ef0dc35696ecbd3e7a7ca7cffebddee9ed"} Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.798963 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda"} Feb 20 07:48:39 crc kubenswrapper[4492]: I0220 07:48:39.799333 4492 scope.go:117] "RemoveContainer" containerID="f04b2dc040e6054f4662cd41991a878606497eb39489c91543e3bb155da42165" Feb 20 07:48:40 crc kubenswrapper[4492]: I0220 07:48:40.579113 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-67rg8" podUID="922456bc-c937-4182-b66c-9b6cc71f80a4" containerName="registry-server" probeResult="failure" output=< Feb 20 07:48:40 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:48:40 crc kubenswrapper[4492]: > Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.624915 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2km7z"] Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.631552 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.679176 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2km7z"] Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.682816 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-utilities\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.683029 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-catalog-content\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.683147 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twpvz\" (UniqueName: \"kubernetes.io/projected/134a2c90-8690-42ac-b4e1-7428364ec4de-kube-api-access-twpvz\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.785337 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-catalog-content\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.785485 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twpvz\" (UniqueName: \"kubernetes.io/projected/134a2c90-8690-42ac-b4e1-7428364ec4de-kube-api-access-twpvz\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.785590 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-utilities\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.786066 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-catalog-content\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.786117 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-utilities\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.807553 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twpvz\" (UniqueName: \"kubernetes.io/projected/134a2c90-8690-42ac-b4e1-7428364ec4de-kube-api-access-twpvz\") pod \"certified-operators-2km7z\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:42 crc kubenswrapper[4492]: I0220 07:48:42.948537 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:43 crc kubenswrapper[4492]: I0220 07:48:43.611088 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2km7z"] Feb 20 07:48:43 crc kubenswrapper[4492]: I0220 07:48:43.835780 4492 generic.go:334] "Generic (PLEG): container finished" podID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerID="a7c88e99fb4d9e161324210f4d0f3c908b50763e1dfeaea99634e450a066d0a0" exitCode=0 Feb 20 07:48:43 crc kubenswrapper[4492]: I0220 07:48:43.835839 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2km7z" event={"ID":"134a2c90-8690-42ac-b4e1-7428364ec4de","Type":"ContainerDied","Data":"a7c88e99fb4d9e161324210f4d0f3c908b50763e1dfeaea99634e450a066d0a0"} Feb 20 07:48:43 crc kubenswrapper[4492]: I0220 07:48:43.835878 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2km7z" event={"ID":"134a2c90-8690-42ac-b4e1-7428364ec4de","Type":"ContainerStarted","Data":"d1aea5b4756c8a76c2d18e4f1694287ee25f29a29aaae367a473d9bef49b2497"} Feb 20 07:48:44 crc kubenswrapper[4492]: I0220 07:48:44.845257 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2km7z" event={"ID":"134a2c90-8690-42ac-b4e1-7428364ec4de","Type":"ContainerStarted","Data":"74dbec0b1db19abdf3d8cfc8d60ac14f8856a4d0b522f34ab3e8ca1e72cac2bf"} Feb 20 07:48:45 crc kubenswrapper[4492]: I0220 07:48:45.854925 4492 generic.go:334] "Generic (PLEG): container finished" podID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerID="74dbec0b1db19abdf3d8cfc8d60ac14f8856a4d0b522f34ab3e8ca1e72cac2bf" exitCode=0 Feb 20 07:48:45 crc kubenswrapper[4492]: I0220 07:48:45.854985 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2km7z" event={"ID":"134a2c90-8690-42ac-b4e1-7428364ec4de","Type":"ContainerDied","Data":"74dbec0b1db19abdf3d8cfc8d60ac14f8856a4d0b522f34ab3e8ca1e72cac2bf"} Feb 20 07:48:46 crc kubenswrapper[4492]: I0220 07:48:46.865978 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2km7z" event={"ID":"134a2c90-8690-42ac-b4e1-7428364ec4de","Type":"ContainerStarted","Data":"932f4e2b40a19e729ad7d9b3c4bffa2acf27c6b81b42e1c0fce0dd409dc4311c"} Feb 20 07:48:49 crc kubenswrapper[4492]: I0220 07:48:49.573069 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:49 crc kubenswrapper[4492]: I0220 07:48:49.588950 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2km7z" podStartSLOduration=5.048005114 podStartE2EDuration="7.588925459s" podCreationTimestamp="2026-02-20 07:48:42 +0000 UTC" firstStartedPulling="2026-02-20 07:48:43.838875138 +0000 UTC m=+4080.610164116" lastFinishedPulling="2026-02-20 07:48:46.379795483 +0000 UTC m=+4083.151084461" observedRunningTime="2026-02-20 07:48:46.886621979 +0000 UTC m=+4083.657910957" watchObservedRunningTime="2026-02-20 07:48:49.588925459 +0000 UTC m=+4086.360214437" Feb 20 07:48:49 crc kubenswrapper[4492]: I0220 07:48:49.613092 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-67rg8" Feb 20 07:48:52 crc kubenswrapper[4492]: I0220 07:48:52.949426 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:52 crc kubenswrapper[4492]: I0220 07:48:52.950055 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:52 crc kubenswrapper[4492]: I0220 07:48:52.994054 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:53 crc kubenswrapper[4492]: I0220 07:48:53.230794 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-67rg8"] Feb 20 07:48:53 crc kubenswrapper[4492]: I0220 07:48:53.965905 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.004327 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8fwhv"] Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.005618 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8fwhv" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="registry-server" containerID="cri-o://2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05" gracePeriod=2 Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.555583 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.757236 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7dkv\" (UniqueName: \"kubernetes.io/projected/c83fd1dd-24ef-45f8-a3ff-264d74862d94-kube-api-access-b7dkv\") pod \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.757294 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-catalog-content\") pod \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.757361 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-utilities\") pod \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\" (UID: \"c83fd1dd-24ef-45f8-a3ff-264d74862d94\") " Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.759525 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-utilities" (OuterVolumeSpecName: "utilities") pod "c83fd1dd-24ef-45f8-a3ff-264d74862d94" (UID: "c83fd1dd-24ef-45f8-a3ff-264d74862d94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.768858 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c83fd1dd-24ef-45f8-a3ff-264d74862d94-kube-api-access-b7dkv" (OuterVolumeSpecName: "kube-api-access-b7dkv") pod "c83fd1dd-24ef-45f8-a3ff-264d74862d94" (UID: "c83fd1dd-24ef-45f8-a3ff-264d74862d94"). InnerVolumeSpecName "kube-api-access-b7dkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.827387 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c83fd1dd-24ef-45f8-a3ff-264d74862d94" (UID: "c83fd1dd-24ef-45f8-a3ff-264d74862d94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.860586 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7dkv\" (UniqueName: \"kubernetes.io/projected/c83fd1dd-24ef-45f8-a3ff-264d74862d94-kube-api-access-b7dkv\") on node \"crc\" DevicePath \"\"" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.860623 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.860634 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c83fd1dd-24ef-45f8-a3ff-264d74862d94-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.938235 4492 generic.go:334] "Generic (PLEG): container finished" podID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerID="2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05" exitCode=0 Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.938372 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fwhv" event={"ID":"c83fd1dd-24ef-45f8-a3ff-264d74862d94","Type":"ContainerDied","Data":"2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05"} Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.938533 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fwhv" event={"ID":"c83fd1dd-24ef-45f8-a3ff-264d74862d94","Type":"ContainerDied","Data":"565a9d7c3c86b616208861d203cd85842e261bf3822b88ae16df4d9653bebb95"} Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.938622 4492 scope.go:117] "RemoveContainer" containerID="2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.938739 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fwhv" Feb 20 07:48:54 crc kubenswrapper[4492]: I0220 07:48:54.991189 4492 scope.go:117] "RemoveContainer" containerID="5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:54.998942 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8fwhv"] Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.006869 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8fwhv"] Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.029047 4492 scope.go:117] "RemoveContainer" containerID="2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.061634 4492 scope.go:117] "RemoveContainer" containerID="2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05" Feb 20 07:48:55 crc kubenswrapper[4492]: E0220 07:48:55.063586 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05\": container with ID starting with 2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05 not found: ID does not exist" containerID="2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.064196 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05"} err="failed to get container status \"2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05\": rpc error: code = NotFound desc = could not find container \"2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05\": container with ID starting with 2c1c20b4c6b0d11ed2a68a3d8201faaef5966630648012712be38c29a5eabd05 not found: ID does not exist" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.064234 4492 scope.go:117] "RemoveContainer" containerID="5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9" Feb 20 07:48:55 crc kubenswrapper[4492]: E0220 07:48:55.064871 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9\": container with ID starting with 5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9 not found: ID does not exist" containerID="5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.065425 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9"} err="failed to get container status \"5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9\": rpc error: code = NotFound desc = could not find container \"5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9\": container with ID starting with 5ba467caf768995d43ddde4b98a315e7a6232c704209c43b378cd7a39d26efe9 not found: ID does not exist" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.065540 4492 scope.go:117] "RemoveContainer" containerID="2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c" Feb 20 07:48:55 crc kubenswrapper[4492]: E0220 07:48:55.066077 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c\": container with ID starting with 2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c not found: ID does not exist" containerID="2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.066108 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c"} err="failed to get container status \"2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c\": rpc error: code = NotFound desc = could not find container \"2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c\": container with ID starting with 2f3d67a4860b59cc40a032e4778630c2f36b303ee9d2eb4ebc06203401f4fa9c not found: ID does not exist" Feb 20 07:48:55 crc kubenswrapper[4492]: I0220 07:48:55.583896 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" path="/var/lib/kubelet/pods/c83fd1dd-24ef-45f8-a3ff-264d74862d94/volumes" Feb 20 07:48:57 crc kubenswrapper[4492]: I0220 07:48:57.790164 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2km7z"] Feb 20 07:48:57 crc kubenswrapper[4492]: I0220 07:48:57.790797 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2km7z" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="registry-server" containerID="cri-o://932f4e2b40a19e729ad7d9b3c4bffa2acf27c6b81b42e1c0fce0dd409dc4311c" gracePeriod=2 Feb 20 07:48:57 crc kubenswrapper[4492]: I0220 07:48:57.968682 4492 generic.go:334] "Generic (PLEG): container finished" podID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerID="932f4e2b40a19e729ad7d9b3c4bffa2acf27c6b81b42e1c0fce0dd409dc4311c" exitCode=0 Feb 20 07:48:57 crc kubenswrapper[4492]: I0220 07:48:57.968731 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2km7z" event={"ID":"134a2c90-8690-42ac-b4e1-7428364ec4de","Type":"ContainerDied","Data":"932f4e2b40a19e729ad7d9b3c4bffa2acf27c6b81b42e1c0fce0dd409dc4311c"} Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.241501 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.246398 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-utilities\") pod \"134a2c90-8690-42ac-b4e1-7428364ec4de\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.246457 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-catalog-content\") pod \"134a2c90-8690-42ac-b4e1-7428364ec4de\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.246684 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twpvz\" (UniqueName: \"kubernetes.io/projected/134a2c90-8690-42ac-b4e1-7428364ec4de-kube-api-access-twpvz\") pod \"134a2c90-8690-42ac-b4e1-7428364ec4de\" (UID: \"134a2c90-8690-42ac-b4e1-7428364ec4de\") " Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.247621 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-utilities" (OuterVolumeSpecName: "utilities") pod "134a2c90-8690-42ac-b4e1-7428364ec4de" (UID: "134a2c90-8690-42ac-b4e1-7428364ec4de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.253150 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/134a2c90-8690-42ac-b4e1-7428364ec4de-kube-api-access-twpvz" (OuterVolumeSpecName: "kube-api-access-twpvz") pod "134a2c90-8690-42ac-b4e1-7428364ec4de" (UID: "134a2c90-8690-42ac-b4e1-7428364ec4de"). InnerVolumeSpecName "kube-api-access-twpvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.301208 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "134a2c90-8690-42ac-b4e1-7428364ec4de" (UID: "134a2c90-8690-42ac-b4e1-7428364ec4de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.349453 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.349515 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/134a2c90-8690-42ac-b4e1-7428364ec4de-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.349530 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twpvz\" (UniqueName: \"kubernetes.io/projected/134a2c90-8690-42ac-b4e1-7428364ec4de-kube-api-access-twpvz\") on node \"crc\" DevicePath \"\"" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.979642 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2km7z" event={"ID":"134a2c90-8690-42ac-b4e1-7428364ec4de","Type":"ContainerDied","Data":"d1aea5b4756c8a76c2d18e4f1694287ee25f29a29aaae367a473d9bef49b2497"} Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.980283 4492 scope.go:117] "RemoveContainer" containerID="932f4e2b40a19e729ad7d9b3c4bffa2acf27c6b81b42e1c0fce0dd409dc4311c" Feb 20 07:48:58 crc kubenswrapper[4492]: I0220 07:48:58.980463 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2km7z" Feb 20 07:48:59 crc kubenswrapper[4492]: I0220 07:48:59.002327 4492 scope.go:117] "RemoveContainer" containerID="74dbec0b1db19abdf3d8cfc8d60ac14f8856a4d0b522f34ab3e8ca1e72cac2bf" Feb 20 07:48:59 crc kubenswrapper[4492]: I0220 07:48:59.023746 4492 scope.go:117] "RemoveContainer" containerID="a7c88e99fb4d9e161324210f4d0f3c908b50763e1dfeaea99634e450a066d0a0" Feb 20 07:48:59 crc kubenswrapper[4492]: I0220 07:48:59.024359 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2km7z"] Feb 20 07:48:59 crc kubenswrapper[4492]: I0220 07:48:59.031390 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2km7z"] Feb 20 07:48:59 crc kubenswrapper[4492]: I0220 07:48:59.577659 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" path="/var/lib/kubelet/pods/134a2c90-8690-42ac-b4e1-7428364ec4de/volumes" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.148701 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rrlg9"] Feb 20 07:49:41 crc kubenswrapper[4492]: E0220 07:49:41.152817 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="extract-utilities" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.152856 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="extract-utilities" Feb 20 07:49:41 crc kubenswrapper[4492]: E0220 07:49:41.152892 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="extract-content" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.152899 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="extract-content" Feb 20 07:49:41 crc kubenswrapper[4492]: E0220 07:49:41.152917 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="extract-utilities" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.152924 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="extract-utilities" Feb 20 07:49:41 crc kubenswrapper[4492]: E0220 07:49:41.152934 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="extract-content" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.152940 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="extract-content" Feb 20 07:49:41 crc kubenswrapper[4492]: E0220 07:49:41.152950 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="registry-server" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.152955 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="registry-server" Feb 20 07:49:41 crc kubenswrapper[4492]: E0220 07:49:41.152979 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="registry-server" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.152986 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="registry-server" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.153707 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c83fd1dd-24ef-45f8-a3ff-264d74862d94" containerName="registry-server" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.153886 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="134a2c90-8690-42ac-b4e1-7428364ec4de" containerName="registry-server" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.155242 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.159539 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrlg9"] Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.195281 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-utilities\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.195373 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-catalog-content\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.195390 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnf7k\" (UniqueName: \"kubernetes.io/projected/c575de10-e0f0-45ae-accf-b92800652617-kube-api-access-jnf7k\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.297255 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-catalog-content\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.297294 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnf7k\" (UniqueName: \"kubernetes.io/projected/c575de10-e0f0-45ae-accf-b92800652617-kube-api-access-jnf7k\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.297531 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-utilities\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.299688 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-catalog-content\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.300221 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-utilities\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.328570 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnf7k\" (UniqueName: \"kubernetes.io/projected/c575de10-e0f0-45ae-accf-b92800652617-kube-api-access-jnf7k\") pod \"redhat-operators-rrlg9\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:41 crc kubenswrapper[4492]: I0220 07:49:41.474902 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:42 crc kubenswrapper[4492]: I0220 07:49:42.032100 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rrlg9"] Feb 20 07:49:42 crc kubenswrapper[4492]: I0220 07:49:42.299944 4492 generic.go:334] "Generic (PLEG): container finished" podID="c575de10-e0f0-45ae-accf-b92800652617" containerID="9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce" exitCode=0 Feb 20 07:49:42 crc kubenswrapper[4492]: I0220 07:49:42.300126 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrlg9" event={"ID":"c575de10-e0f0-45ae-accf-b92800652617","Type":"ContainerDied","Data":"9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce"} Feb 20 07:49:42 crc kubenswrapper[4492]: I0220 07:49:42.300808 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrlg9" event={"ID":"c575de10-e0f0-45ae-accf-b92800652617","Type":"ContainerStarted","Data":"b21a350dcd103f9990e06e398ffbb19ab74ba3a3913ca57cdc90bee9e7be5d3c"} Feb 20 07:49:43 crc kubenswrapper[4492]: I0220 07:49:43.309566 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrlg9" event={"ID":"c575de10-e0f0-45ae-accf-b92800652617","Type":"ContainerStarted","Data":"0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93"} Feb 20 07:49:46 crc kubenswrapper[4492]: I0220 07:49:46.335093 4492 generic.go:334] "Generic (PLEG): container finished" podID="c575de10-e0f0-45ae-accf-b92800652617" containerID="0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93" exitCode=0 Feb 20 07:49:46 crc kubenswrapper[4492]: I0220 07:49:46.335189 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrlg9" event={"ID":"c575de10-e0f0-45ae-accf-b92800652617","Type":"ContainerDied","Data":"0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93"} Feb 20 07:49:47 crc kubenswrapper[4492]: I0220 07:49:47.345912 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrlg9" event={"ID":"c575de10-e0f0-45ae-accf-b92800652617","Type":"ContainerStarted","Data":"5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9"} Feb 20 07:49:47 crc kubenswrapper[4492]: I0220 07:49:47.368902 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rrlg9" podStartSLOduration=1.853291268 podStartE2EDuration="6.368668468s" podCreationTimestamp="2026-02-20 07:49:41 +0000 UTC" firstStartedPulling="2026-02-20 07:49:42.301808509 +0000 UTC m=+4139.073097487" lastFinishedPulling="2026-02-20 07:49:46.817185719 +0000 UTC m=+4143.588474687" observedRunningTime="2026-02-20 07:49:47.365677749 +0000 UTC m=+4144.136966717" watchObservedRunningTime="2026-02-20 07:49:47.368668468 +0000 UTC m=+4144.139957446" Feb 20 07:49:47 crc kubenswrapper[4492]: I0220 07:49:47.954205 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bn8d6"] Feb 20 07:49:47 crc kubenswrapper[4492]: I0220 07:49:47.956286 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:47 crc kubenswrapper[4492]: I0220 07:49:47.966126 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn8d6"] Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.018042 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-utilities\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.018096 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-catalog-content\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.018174 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm6wh\" (UniqueName: \"kubernetes.io/projected/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-kube-api-access-tm6wh\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.120022 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-utilities\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.120106 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-catalog-content\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.120428 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-utilities\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.120532 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-catalog-content\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.120661 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm6wh\" (UniqueName: \"kubernetes.io/projected/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-kube-api-access-tm6wh\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.141673 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm6wh\" (UniqueName: \"kubernetes.io/projected/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-kube-api-access-tm6wh\") pod \"redhat-marketplace-bn8d6\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.271917 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:48 crc kubenswrapper[4492]: I0220 07:49:48.997772 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn8d6"] Feb 20 07:49:49 crc kubenswrapper[4492]: I0220 07:49:49.366353 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerStarted","Data":"afd06598c7f1ae38d825db66e55f294b2ae520c27e07b3572971336989ed8b7f"} Feb 20 07:49:49 crc kubenswrapper[4492]: I0220 07:49:49.366394 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerStarted","Data":"f6ab1b8bf17355f3f8d1b4119f42e33779b37b08a62f4563df46f6487e02ab3d"} Feb 20 07:49:50 crc kubenswrapper[4492]: I0220 07:49:50.377626 4492 generic.go:334] "Generic (PLEG): container finished" podID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerID="afd06598c7f1ae38d825db66e55f294b2ae520c27e07b3572971336989ed8b7f" exitCode=0 Feb 20 07:49:50 crc kubenswrapper[4492]: I0220 07:49:50.377687 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerDied","Data":"afd06598c7f1ae38d825db66e55f294b2ae520c27e07b3572971336989ed8b7f"} Feb 20 07:49:50 crc kubenswrapper[4492]: I0220 07:49:50.377919 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerStarted","Data":"5b18067678f2ad1de2285bbe8645229201335736d608ee638d80b502df79f99b"} Feb 20 07:49:51 crc kubenswrapper[4492]: I0220 07:49:51.393844 4492 generic.go:334] "Generic (PLEG): container finished" podID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerID="5b18067678f2ad1de2285bbe8645229201335736d608ee638d80b502df79f99b" exitCode=0 Feb 20 07:49:51 crc kubenswrapper[4492]: I0220 07:49:51.393936 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerDied","Data":"5b18067678f2ad1de2285bbe8645229201335736d608ee638d80b502df79f99b"} Feb 20 07:49:51 crc kubenswrapper[4492]: I0220 07:49:51.475352 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:51 crc kubenswrapper[4492]: I0220 07:49:51.475563 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:49:52 crc kubenswrapper[4492]: I0220 07:49:52.406687 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerStarted","Data":"477c4669912cd69908b3f1fce9ab55eacaf03406d5856a8a176419484f2ff439"} Feb 20 07:49:52 crc kubenswrapper[4492]: I0220 07:49:52.424077 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bn8d6" podStartSLOduration=2.900748352 podStartE2EDuration="5.424053168s" podCreationTimestamp="2026-02-20 07:49:47 +0000 UTC" firstStartedPulling="2026-02-20 07:49:49.368280461 +0000 UTC m=+4146.139569429" lastFinishedPulling="2026-02-20 07:49:51.891585267 +0000 UTC m=+4148.662874245" observedRunningTime="2026-02-20 07:49:52.42376772 +0000 UTC m=+4149.195056698" watchObservedRunningTime="2026-02-20 07:49:52.424053168 +0000 UTC m=+4149.195342146" Feb 20 07:49:52 crc kubenswrapper[4492]: I0220 07:49:52.520890 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrlg9" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="registry-server" probeResult="failure" output=< Feb 20 07:49:52 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:49:52 crc kubenswrapper[4492]: > Feb 20 07:49:58 crc kubenswrapper[4492]: I0220 07:49:58.273054 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:58 crc kubenswrapper[4492]: I0220 07:49:58.273750 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:58 crc kubenswrapper[4492]: I0220 07:49:58.313230 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:58 crc kubenswrapper[4492]: I0220 07:49:58.516255 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:49:58 crc kubenswrapper[4492]: I0220 07:49:58.562904 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn8d6"] Feb 20 07:50:00 crc kubenswrapper[4492]: I0220 07:50:00.500267 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bn8d6" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="registry-server" containerID="cri-o://477c4669912cd69908b3f1fce9ab55eacaf03406d5856a8a176419484f2ff439" gracePeriod=2 Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.508090 4492 generic.go:334] "Generic (PLEG): container finished" podID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerID="477c4669912cd69908b3f1fce9ab55eacaf03406d5856a8a176419484f2ff439" exitCode=0 Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.508639 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerDied","Data":"477c4669912cd69908b3f1fce9ab55eacaf03406d5856a8a176419484f2ff439"} Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.700277 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.832676 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm6wh\" (UniqueName: \"kubernetes.io/projected/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-kube-api-access-tm6wh\") pod \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.833146 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-utilities\") pod \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.833287 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-catalog-content\") pod \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\" (UID: \"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a\") " Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.834897 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-utilities" (OuterVolumeSpecName: "utilities") pod "2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" (UID: "2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.843802 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" (UID: "2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.910955 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-kube-api-access-tm6wh" (OuterVolumeSpecName: "kube-api-access-tm6wh") pod "2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" (UID: "2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a"). InnerVolumeSpecName "kube-api-access-tm6wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.936526 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm6wh\" (UniqueName: \"kubernetes.io/projected/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-kube-api-access-tm6wh\") on node \"crc\" DevicePath \"\"" Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.936553 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:50:01 crc kubenswrapper[4492]: I0220 07:50:01.936563 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.518334 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bn8d6" event={"ID":"2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a","Type":"ContainerDied","Data":"f6ab1b8bf17355f3f8d1b4119f42e33779b37b08a62f4563df46f6487e02ab3d"} Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.518460 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bn8d6" Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.520169 4492 scope.go:117] "RemoveContainer" containerID="477c4669912cd69908b3f1fce9ab55eacaf03406d5856a8a176419484f2ff439" Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.553460 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn8d6"] Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.555192 4492 scope.go:117] "RemoveContainer" containerID="5b18067678f2ad1de2285bbe8645229201335736d608ee638d80b502df79f99b" Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.559449 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bn8d6"] Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.572752 4492 scope.go:117] "RemoveContainer" containerID="afd06598c7f1ae38d825db66e55f294b2ae520c27e07b3572971336989ed8b7f" Feb 20 07:50:02 crc kubenswrapper[4492]: I0220 07:50:02.641289 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rrlg9" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="registry-server" probeResult="failure" output=< Feb 20 07:50:02 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:50:02 crc kubenswrapper[4492]: > Feb 20 07:50:03 crc kubenswrapper[4492]: I0220 07:50:03.580365 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" path="/var/lib/kubelet/pods/2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a/volumes" Feb 20 07:50:11 crc kubenswrapper[4492]: I0220 07:50:11.737136 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:50:11 crc kubenswrapper[4492]: I0220 07:50:11.776369 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:50:12 crc kubenswrapper[4492]: I0220 07:50:12.352115 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrlg9"] Feb 20 07:50:13 crc kubenswrapper[4492]: I0220 07:50:13.597388 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rrlg9" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="registry-server" containerID="cri-o://5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9" gracePeriod=2 Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.062070 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.183280 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-utilities\") pod \"c575de10-e0f0-45ae-accf-b92800652617\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.183384 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-catalog-content\") pod \"c575de10-e0f0-45ae-accf-b92800652617\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.183408 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnf7k\" (UniqueName: \"kubernetes.io/projected/c575de10-e0f0-45ae-accf-b92800652617-kube-api-access-jnf7k\") pod \"c575de10-e0f0-45ae-accf-b92800652617\" (UID: \"c575de10-e0f0-45ae-accf-b92800652617\") " Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.185149 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-utilities" (OuterVolumeSpecName: "utilities") pod "c575de10-e0f0-45ae-accf-b92800652617" (UID: "c575de10-e0f0-45ae-accf-b92800652617"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.192716 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c575de10-e0f0-45ae-accf-b92800652617-kube-api-access-jnf7k" (OuterVolumeSpecName: "kube-api-access-jnf7k") pod "c575de10-e0f0-45ae-accf-b92800652617" (UID: "c575de10-e0f0-45ae-accf-b92800652617"). InnerVolumeSpecName "kube-api-access-jnf7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.286700 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnf7k\" (UniqueName: \"kubernetes.io/projected/c575de10-e0f0-45ae-accf-b92800652617-kube-api-access-jnf7k\") on node \"crc\" DevicePath \"\"" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.286728 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.309188 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c575de10-e0f0-45ae-accf-b92800652617" (UID: "c575de10-e0f0-45ae-accf-b92800652617"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.388525 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c575de10-e0f0-45ae-accf-b92800652617-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.607020 4492 generic.go:334] "Generic (PLEG): container finished" podID="c575de10-e0f0-45ae-accf-b92800652617" containerID="5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9" exitCode=0 Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.607056 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrlg9" event={"ID":"c575de10-e0f0-45ae-accf-b92800652617","Type":"ContainerDied","Data":"5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9"} Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.607082 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rrlg9" event={"ID":"c575de10-e0f0-45ae-accf-b92800652617","Type":"ContainerDied","Data":"b21a350dcd103f9990e06e398ffbb19ab74ba3a3913ca57cdc90bee9e7be5d3c"} Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.607101 4492 scope.go:117] "RemoveContainer" containerID="5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.607123 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rrlg9" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.627441 4492 scope.go:117] "RemoveContainer" containerID="0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.639604 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rrlg9"] Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.646285 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rrlg9"] Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.647137 4492 scope.go:117] "RemoveContainer" containerID="9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.685805 4492 scope.go:117] "RemoveContainer" containerID="5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9" Feb 20 07:50:14 crc kubenswrapper[4492]: E0220 07:50:14.687599 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9\": container with ID starting with 5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9 not found: ID does not exist" containerID="5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.687646 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9"} err="failed to get container status \"5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9\": rpc error: code = NotFound desc = could not find container \"5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9\": container with ID starting with 5d29f035b429228372fd396fa3c43be01282777ef9c80d90e92cc775d1f172c9 not found: ID does not exist" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.687667 4492 scope.go:117] "RemoveContainer" containerID="0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93" Feb 20 07:50:14 crc kubenswrapper[4492]: E0220 07:50:14.688206 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93\": container with ID starting with 0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93 not found: ID does not exist" containerID="0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.688273 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93"} err="failed to get container status \"0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93\": rpc error: code = NotFound desc = could not find container \"0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93\": container with ID starting with 0f243e94d04f10e00bc246026015cf9c926cdd8f19b5dd90da4c9bae5011ea93 not found: ID does not exist" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.688309 4492 scope.go:117] "RemoveContainer" containerID="9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce" Feb 20 07:50:14 crc kubenswrapper[4492]: E0220 07:50:14.688833 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce\": container with ID starting with 9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce not found: ID does not exist" containerID="9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce" Feb 20 07:50:14 crc kubenswrapper[4492]: I0220 07:50:14.688861 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce"} err="failed to get container status \"9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce\": rpc error: code = NotFound desc = could not find container \"9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce\": container with ID starting with 9beb5b896dca0c187b8d49ea79c3d7c99278df41bf767d7f4d302a4e41d3c3ce not found: ID does not exist" Feb 20 07:50:15 crc kubenswrapper[4492]: I0220 07:50:15.566843 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c575de10-e0f0-45ae-accf-b92800652617" path="/var/lib/kubelet/pods/c575de10-e0f0-45ae-accf-b92800652617/volumes" Feb 20 07:50:39 crc kubenswrapper[4492]: I0220 07:50:39.311418 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:50:39 crc kubenswrapper[4492]: I0220 07:50:39.311869 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:51:09 crc kubenswrapper[4492]: I0220 07:51:09.311911 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:51:09 crc kubenswrapper[4492]: I0220 07:51:09.312741 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:51:39 crc kubenswrapper[4492]: I0220 07:51:39.311638 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:51:39 crc kubenswrapper[4492]: I0220 07:51:39.312189 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:51:39 crc kubenswrapper[4492]: I0220 07:51:39.313019 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 07:51:39 crc kubenswrapper[4492]: I0220 07:51:39.313457 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 07:51:39 crc kubenswrapper[4492]: I0220 07:51:39.313531 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" gracePeriod=600 Feb 20 07:51:39 crc kubenswrapper[4492]: E0220 07:51:39.443964 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:51:40 crc kubenswrapper[4492]: I0220 07:51:40.304379 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" exitCode=0 Feb 20 07:51:40 crc kubenswrapper[4492]: I0220 07:51:40.304425 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda"} Feb 20 07:51:40 crc kubenswrapper[4492]: I0220 07:51:40.304485 4492 scope.go:117] "RemoveContainer" containerID="4df0d1375018eccda77abe46780f42ef0dc35696ecbd3e7a7ca7cffebddee9ed" Feb 20 07:51:40 crc kubenswrapper[4492]: I0220 07:51:40.306338 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:51:40 crc kubenswrapper[4492]: E0220 07:51:40.306899 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:51:54 crc kubenswrapper[4492]: I0220 07:51:54.557307 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:51:54 crc kubenswrapper[4492]: E0220 07:51:54.558025 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:52:07 crc kubenswrapper[4492]: I0220 07:52:07.557231 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:52:07 crc kubenswrapper[4492]: E0220 07:52:07.558208 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:52:20 crc kubenswrapper[4492]: I0220 07:52:20.556824 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:52:20 crc kubenswrapper[4492]: E0220 07:52:20.557328 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:52:31 crc kubenswrapper[4492]: I0220 07:52:31.556952 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:52:31 crc kubenswrapper[4492]: E0220 07:52:31.557464 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:52:45 crc kubenswrapper[4492]: I0220 07:52:45.557112 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:52:45 crc kubenswrapper[4492]: E0220 07:52:45.557692 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:53:00 crc kubenswrapper[4492]: I0220 07:53:00.556557 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:53:00 crc kubenswrapper[4492]: E0220 07:53:00.557223 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:53:13 crc kubenswrapper[4492]: I0220 07:53:13.561632 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:53:13 crc kubenswrapper[4492]: E0220 07:53:13.562280 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:53:26 crc kubenswrapper[4492]: I0220 07:53:26.557609 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:53:26 crc kubenswrapper[4492]: E0220 07:53:26.558458 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:53:37 crc kubenswrapper[4492]: I0220 07:53:37.557648 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:53:37 crc kubenswrapper[4492]: E0220 07:53:37.558526 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:53:52 crc kubenswrapper[4492]: I0220 07:53:52.557063 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:53:52 crc kubenswrapper[4492]: E0220 07:53:52.557767 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:54:04 crc kubenswrapper[4492]: I0220 07:54:04.557284 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:54:04 crc kubenswrapper[4492]: E0220 07:54:04.557818 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:54:19 crc kubenswrapper[4492]: I0220 07:54:19.557942 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:54:19 crc kubenswrapper[4492]: E0220 07:54:19.558457 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:54:33 crc kubenswrapper[4492]: I0220 07:54:33.561927 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:54:33 crc kubenswrapper[4492]: E0220 07:54:33.562671 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:54:47 crc kubenswrapper[4492]: I0220 07:54:47.557788 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:54:47 crc kubenswrapper[4492]: E0220 07:54:47.558615 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:55:01 crc kubenswrapper[4492]: I0220 07:55:01.556691 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:55:01 crc kubenswrapper[4492]: E0220 07:55:01.557497 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:55:13 crc kubenswrapper[4492]: I0220 07:55:13.561642 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:55:13 crc kubenswrapper[4492]: E0220 07:55:13.563672 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:55:27 crc kubenswrapper[4492]: I0220 07:55:27.557057 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:55:27 crc kubenswrapper[4492]: E0220 07:55:27.557952 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:55:40 crc kubenswrapper[4492]: I0220 07:55:40.556656 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:55:40 crc kubenswrapper[4492]: E0220 07:55:40.557377 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:55:51 crc kubenswrapper[4492]: I0220 07:55:51.557348 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:55:51 crc kubenswrapper[4492]: E0220 07:55:51.559028 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:56:05 crc kubenswrapper[4492]: I0220 07:56:05.557896 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:56:05 crc kubenswrapper[4492]: E0220 07:56:05.559398 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:56:19 crc kubenswrapper[4492]: I0220 07:56:19.563531 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:56:19 crc kubenswrapper[4492]: E0220 07:56:19.566218 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:56:33 crc kubenswrapper[4492]: I0220 07:56:33.562087 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:56:33 crc kubenswrapper[4492]: E0220 07:56:33.562952 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 07:56:47 crc kubenswrapper[4492]: I0220 07:56:47.557221 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 07:56:48 crc kubenswrapper[4492]: I0220 07:56:48.509557 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"046d2f99be3935b7577536a5396fd26213ad42cda70983181ddfca7d9b35ec0b"} Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.981709 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hgdjq"] Feb 20 07:58:38 crc kubenswrapper[4492]: E0220 07:58:38.983958 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="registry-server" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.983983 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="registry-server" Feb 20 07:58:38 crc kubenswrapper[4492]: E0220 07:58:38.984004 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="extract-content" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.984011 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="extract-content" Feb 20 07:58:38 crc kubenswrapper[4492]: E0220 07:58:38.984053 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="registry-server" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.984059 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="registry-server" Feb 20 07:58:38 crc kubenswrapper[4492]: E0220 07:58:38.984077 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="extract-utilities" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.984083 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="extract-utilities" Feb 20 07:58:38 crc kubenswrapper[4492]: E0220 07:58:38.984094 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="extract-utilities" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.984099 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="extract-utilities" Feb 20 07:58:38 crc kubenswrapper[4492]: E0220 07:58:38.984111 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="extract-content" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.984116 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="extract-content" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.984641 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c575de10-e0f0-45ae-accf-b92800652617" containerName="registry-server" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.984670 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="2809dfcc-d45b-4e1e-a461-0eb28d6b4f7a" containerName="registry-server" Feb 20 07:58:38 crc kubenswrapper[4492]: I0220 07:58:38.989770 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.002237 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hgdjq"] Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.011054 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-catalog-content\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.011198 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-utilities\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.011401 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ddnz\" (UniqueName: \"kubernetes.io/projected/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-kube-api-access-6ddnz\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.114208 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-catalog-content\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.114279 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-utilities\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.114318 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ddnz\" (UniqueName: \"kubernetes.io/projected/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-kube-api-access-6ddnz\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.114794 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-catalog-content\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.114875 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-utilities\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.134619 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ddnz\" (UniqueName: \"kubernetes.io/projected/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-kube-api-access-6ddnz\") pod \"community-operators-hgdjq\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:39 crc kubenswrapper[4492]: I0220 07:58:39.312885 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:40 crc kubenswrapper[4492]: I0220 07:58:40.004049 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hgdjq"] Feb 20 07:58:40 crc kubenswrapper[4492]: I0220 07:58:40.472806 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgdjq" event={"ID":"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0","Type":"ContainerStarted","Data":"67defe4f654f3abc7b67347b27a00128f7cdbffb54d9286946384ecfaefcede9"} Feb 20 07:58:41 crc kubenswrapper[4492]: I0220 07:58:41.485128 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgdjq" event={"ID":"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0","Type":"ContainerDied","Data":"63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446"} Feb 20 07:58:41 crc kubenswrapper[4492]: I0220 07:58:41.485749 4492 generic.go:334] "Generic (PLEG): container finished" podID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerID="63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446" exitCode=0 Feb 20 07:58:41 crc kubenswrapper[4492]: I0220 07:58:41.491296 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 07:58:43 crc kubenswrapper[4492]: I0220 07:58:43.504734 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgdjq" event={"ID":"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0","Type":"ContainerStarted","Data":"0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026"} Feb 20 07:58:44 crc kubenswrapper[4492]: I0220 07:58:44.517859 4492 generic.go:334] "Generic (PLEG): container finished" podID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerID="0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026" exitCode=0 Feb 20 07:58:44 crc kubenswrapper[4492]: I0220 07:58:44.518073 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgdjq" event={"ID":"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0","Type":"ContainerDied","Data":"0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026"} Feb 20 07:58:45 crc kubenswrapper[4492]: I0220 07:58:45.531355 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgdjq" event={"ID":"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0","Type":"ContainerStarted","Data":"e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b"} Feb 20 07:58:45 crc kubenswrapper[4492]: I0220 07:58:45.560355 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hgdjq" podStartSLOduration=4.084328704 podStartE2EDuration="7.555832913s" podCreationTimestamp="2026-02-20 07:58:38 +0000 UTC" firstStartedPulling="2026-02-20 07:58:41.487300106 +0000 UTC m=+4678.258589084" lastFinishedPulling="2026-02-20 07:58:44.958804314 +0000 UTC m=+4681.730093293" observedRunningTime="2026-02-20 07:58:45.550984259 +0000 UTC m=+4682.322273227" watchObservedRunningTime="2026-02-20 07:58:45.555832913 +0000 UTC m=+4682.327121891" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.333739 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xwbn9"] Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.336454 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.350154 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwbn9"] Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.390176 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-utilities\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.390263 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr8kj\" (UniqueName: \"kubernetes.io/projected/b1bddb70-b394-482f-8993-62c877f8b055-kube-api-access-vr8kj\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.390351 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-catalog-content\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.493809 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-utilities\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.493887 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr8kj\" (UniqueName: \"kubernetes.io/projected/b1bddb70-b394-482f-8993-62c877f8b055-kube-api-access-vr8kj\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.493978 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-catalog-content\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.494269 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-utilities\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.494508 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-catalog-content\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.515729 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr8kj\" (UniqueName: \"kubernetes.io/projected/b1bddb70-b394-482f-8993-62c877f8b055-kube-api-access-vr8kj\") pod \"certified-operators-xwbn9\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:46 crc kubenswrapper[4492]: I0220 07:58:46.665404 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:47 crc kubenswrapper[4492]: I0220 07:58:47.215849 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xwbn9"] Feb 20 07:58:47 crc kubenswrapper[4492]: I0220 07:58:47.552158 4492 generic.go:334] "Generic (PLEG): container finished" podID="b1bddb70-b394-482f-8993-62c877f8b055" containerID="f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214" exitCode=0 Feb 20 07:58:47 crc kubenswrapper[4492]: I0220 07:58:47.552214 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwbn9" event={"ID":"b1bddb70-b394-482f-8993-62c877f8b055","Type":"ContainerDied","Data":"f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214"} Feb 20 07:58:47 crc kubenswrapper[4492]: I0220 07:58:47.552442 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwbn9" event={"ID":"b1bddb70-b394-482f-8993-62c877f8b055","Type":"ContainerStarted","Data":"9d93c57d01d06fbfaa8fcc36fbcb2eb81ad81c8d69c6dd6290d9486767f9bbec"} Feb 20 07:58:48 crc kubenswrapper[4492]: I0220 07:58:48.567278 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwbn9" event={"ID":"b1bddb70-b394-482f-8993-62c877f8b055","Type":"ContainerStarted","Data":"4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2"} Feb 20 07:58:49 crc kubenswrapper[4492]: I0220 07:58:49.314077 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:49 crc kubenswrapper[4492]: I0220 07:58:49.314148 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:49 crc kubenswrapper[4492]: I0220 07:58:49.584170 4492 generic.go:334] "Generic (PLEG): container finished" podID="b1bddb70-b394-482f-8993-62c877f8b055" containerID="4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2" exitCode=0 Feb 20 07:58:49 crc kubenswrapper[4492]: I0220 07:58:49.584503 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwbn9" event={"ID":"b1bddb70-b394-482f-8993-62c877f8b055","Type":"ContainerDied","Data":"4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2"} Feb 20 07:58:50 crc kubenswrapper[4492]: E0220 07:58:50.210047 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e54b5b_dac1_4bc4_88b8_aa3558f84bf0.slice/crio-0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026.scope\": RecentStats: unable to find data in memory cache]" Feb 20 07:58:50 crc kubenswrapper[4492]: I0220 07:58:50.403521 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hgdjq" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="registry-server" probeResult="failure" output=< Feb 20 07:58:50 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:58:50 crc kubenswrapper[4492]: > Feb 20 07:58:50 crc kubenswrapper[4492]: I0220 07:58:50.606153 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwbn9" event={"ID":"b1bddb70-b394-482f-8993-62c877f8b055","Type":"ContainerStarted","Data":"119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b"} Feb 20 07:58:50 crc kubenswrapper[4492]: I0220 07:58:50.633158 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xwbn9" podStartSLOduration=2.13910246 podStartE2EDuration="4.633130767s" podCreationTimestamp="2026-02-20 07:58:46 +0000 UTC" firstStartedPulling="2026-02-20 07:58:47.553785888 +0000 UTC m=+4684.325074865" lastFinishedPulling="2026-02-20 07:58:50.047814193 +0000 UTC m=+4686.819103172" observedRunningTime="2026-02-20 07:58:50.627292055 +0000 UTC m=+4687.398581033" watchObservedRunningTime="2026-02-20 07:58:50.633130767 +0000 UTC m=+4687.404419745" Feb 20 07:58:56 crc kubenswrapper[4492]: I0220 07:58:56.665641 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:56 crc kubenswrapper[4492]: I0220 07:58:56.666344 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:56 crc kubenswrapper[4492]: I0220 07:58:56.704724 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:57 crc kubenswrapper[4492]: I0220 07:58:57.716139 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:58:57 crc kubenswrapper[4492]: I0220 07:58:57.768181 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwbn9"] Feb 20 07:58:59 crc kubenswrapper[4492]: I0220 07:58:59.352464 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:59 crc kubenswrapper[4492]: I0220 07:58:59.396248 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:58:59 crc kubenswrapper[4492]: I0220 07:58:59.691255 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xwbn9" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="registry-server" containerID="cri-o://119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b" gracePeriod=2 Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.211989 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.345763 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr8kj\" (UniqueName: \"kubernetes.io/projected/b1bddb70-b394-482f-8993-62c877f8b055-kube-api-access-vr8kj\") pod \"b1bddb70-b394-482f-8993-62c877f8b055\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.346033 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-utilities\") pod \"b1bddb70-b394-482f-8993-62c877f8b055\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.346693 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-catalog-content\") pod \"b1bddb70-b394-482f-8993-62c877f8b055\" (UID: \"b1bddb70-b394-482f-8993-62c877f8b055\") " Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.348983 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-utilities" (OuterVolumeSpecName: "utilities") pod "b1bddb70-b394-482f-8993-62c877f8b055" (UID: "b1bddb70-b394-482f-8993-62c877f8b055"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.377863 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1bddb70-b394-482f-8993-62c877f8b055-kube-api-access-vr8kj" (OuterVolumeSpecName: "kube-api-access-vr8kj") pod "b1bddb70-b394-482f-8993-62c877f8b055" (UID: "b1bddb70-b394-482f-8993-62c877f8b055"). InnerVolumeSpecName "kube-api-access-vr8kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.393573 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hgdjq"] Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.409909 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1bddb70-b394-482f-8993-62c877f8b055" (UID: "b1bddb70-b394-482f-8993-62c877f8b055"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.449916 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr8kj\" (UniqueName: \"kubernetes.io/projected/b1bddb70-b394-482f-8993-62c877f8b055-kube-api-access-vr8kj\") on node \"crc\" DevicePath \"\"" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.449950 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.449962 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bddb70-b394-482f-8993-62c877f8b055-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:59:00 crc kubenswrapper[4492]: E0220 07:59:00.492536 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e54b5b_dac1_4bc4_88b8_aa3558f84bf0.slice/crio-0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026.scope\": RecentStats: unable to find data in memory cache]" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.705953 4492 generic.go:334] "Generic (PLEG): container finished" podID="b1bddb70-b394-482f-8993-62c877f8b055" containerID="119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b" exitCode=0 Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.706458 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hgdjq" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="registry-server" containerID="cri-o://e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b" gracePeriod=2 Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.706044 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwbn9" event={"ID":"b1bddb70-b394-482f-8993-62c877f8b055","Type":"ContainerDied","Data":"119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b"} Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.706594 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xwbn9" event={"ID":"b1bddb70-b394-482f-8993-62c877f8b055","Type":"ContainerDied","Data":"9d93c57d01d06fbfaa8fcc36fbcb2eb81ad81c8d69c6dd6290d9486767f9bbec"} Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.706636 4492 scope.go:117] "RemoveContainer" containerID="119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.706087 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xwbn9" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.748807 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xwbn9"] Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.790801 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xwbn9"] Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.803532 4492 scope.go:117] "RemoveContainer" containerID="4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.895368 4492 scope.go:117] "RemoveContainer" containerID="f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.927994 4492 scope.go:117] "RemoveContainer" containerID="119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b" Feb 20 07:59:00 crc kubenswrapper[4492]: E0220 07:59:00.938580 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b\": container with ID starting with 119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b not found: ID does not exist" containerID="119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.939288 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b"} err="failed to get container status \"119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b\": rpc error: code = NotFound desc = could not find container \"119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b\": container with ID starting with 119f944b3338d0d14ef9222f8960512804d4a63d2607223c87bec7973f514d6b not found: ID does not exist" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.939329 4492 scope.go:117] "RemoveContainer" containerID="4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2" Feb 20 07:59:00 crc kubenswrapper[4492]: E0220 07:59:00.939699 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2\": container with ID starting with 4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2 not found: ID does not exist" containerID="4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.939814 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2"} err="failed to get container status \"4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2\": rpc error: code = NotFound desc = could not find container \"4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2\": container with ID starting with 4f40014ab5a637f37ffed077bdebb90884d0ab1cecbc2bfa94a57c79da6e1aa2 not found: ID does not exist" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.939894 4492 scope.go:117] "RemoveContainer" containerID="f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214" Feb 20 07:59:00 crc kubenswrapper[4492]: E0220 07:59:00.940246 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214\": container with ID starting with f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214 not found: ID does not exist" containerID="f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214" Feb 20 07:59:00 crc kubenswrapper[4492]: I0220 07:59:00.940286 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214"} err="failed to get container status \"f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214\": rpc error: code = NotFound desc = could not find container \"f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214\": container with ID starting with f2d765d985c44034957cdfa9ebf16c45fc11541f8a377355350f0aa02bc8d214 not found: ID does not exist" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.237772 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.268784 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-utilities\") pod \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.268881 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-catalog-content\") pod \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.269120 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ddnz\" (UniqueName: \"kubernetes.io/projected/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-kube-api-access-6ddnz\") pod \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\" (UID: \"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0\") " Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.272839 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-utilities" (OuterVolumeSpecName: "utilities") pod "06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" (UID: "06e54b5b-dac1-4bc4-88b8-aa3558f84bf0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.277081 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-kube-api-access-6ddnz" (OuterVolumeSpecName: "kube-api-access-6ddnz") pod "06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" (UID: "06e54b5b-dac1-4bc4-88b8-aa3558f84bf0"). InnerVolumeSpecName "kube-api-access-6ddnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.318555 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" (UID: "06e54b5b-dac1-4bc4-88b8-aa3558f84bf0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.371670 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.371705 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.371721 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ddnz\" (UniqueName: \"kubernetes.io/projected/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0-kube-api-access-6ddnz\") on node \"crc\" DevicePath \"\"" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.569140 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1bddb70-b394-482f-8993-62c877f8b055" path="/var/lib/kubelet/pods/b1bddb70-b394-482f-8993-62c877f8b055/volumes" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.717388 4492 generic.go:334] "Generic (PLEG): container finished" podID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerID="e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b" exitCode=0 Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.717435 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgdjq" event={"ID":"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0","Type":"ContainerDied","Data":"e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b"} Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.717507 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hgdjq" event={"ID":"06e54b5b-dac1-4bc4-88b8-aa3558f84bf0","Type":"ContainerDied","Data":"67defe4f654f3abc7b67347b27a00128f7cdbffb54d9286946384ecfaefcede9"} Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.717532 4492 scope.go:117] "RemoveContainer" containerID="e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.717697 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hgdjq" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.744016 4492 scope.go:117] "RemoveContainer" containerID="0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.748111 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hgdjq"] Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.761189 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hgdjq"] Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.765577 4492 scope.go:117] "RemoveContainer" containerID="63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.787926 4492 scope.go:117] "RemoveContainer" containerID="e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b" Feb 20 07:59:01 crc kubenswrapper[4492]: E0220 07:59:01.788388 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b\": container with ID starting with e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b not found: ID does not exist" containerID="e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.788446 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b"} err="failed to get container status \"e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b\": rpc error: code = NotFound desc = could not find container \"e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b\": container with ID starting with e82338a9476c846951534e322899fc4b4b93dc6283bd9c11a68bf8a92bed1d1b not found: ID does not exist" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.788999 4492 scope.go:117] "RemoveContainer" containerID="0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026" Feb 20 07:59:01 crc kubenswrapper[4492]: E0220 07:59:01.789375 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026\": container with ID starting with 0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026 not found: ID does not exist" containerID="0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.789404 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026"} err="failed to get container status \"0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026\": rpc error: code = NotFound desc = could not find container \"0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026\": container with ID starting with 0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026 not found: ID does not exist" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.789428 4492 scope.go:117] "RemoveContainer" containerID="63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446" Feb 20 07:59:01 crc kubenswrapper[4492]: E0220 07:59:01.789720 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446\": container with ID starting with 63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446 not found: ID does not exist" containerID="63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446" Feb 20 07:59:01 crc kubenswrapper[4492]: I0220 07:59:01.789750 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446"} err="failed to get container status \"63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446\": rpc error: code = NotFound desc = could not find container \"63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446\": container with ID starting with 63fe9c74c92cb7c3ba6ebadbb5a88291f3b05ea14fd687f2f4292e20e3279446 not found: ID does not exist" Feb 20 07:59:03 crc kubenswrapper[4492]: I0220 07:59:03.567994 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" path="/var/lib/kubelet/pods/06e54b5b-dac1-4bc4-88b8-aa3558f84bf0/volumes" Feb 20 07:59:09 crc kubenswrapper[4492]: I0220 07:59:09.310975 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:59:09 crc kubenswrapper[4492]: I0220 07:59:09.312250 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:59:10 crc kubenswrapper[4492]: E0220 07:59:10.715845 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e54b5b_dac1_4bc4_88b8_aa3558f84bf0.slice/crio-0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026.scope\": RecentStats: unable to find data in memory cache]" Feb 20 07:59:20 crc kubenswrapper[4492]: E0220 07:59:20.899142 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e54b5b_dac1_4bc4_88b8_aa3558f84bf0.slice/crio-0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026.scope\": RecentStats: unable to find data in memory cache]" Feb 20 07:59:31 crc kubenswrapper[4492]: E0220 07:59:31.093867 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e54b5b_dac1_4bc4_88b8_aa3558f84bf0.slice/crio-0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026.scope\": RecentStats: unable to find data in memory cache]" Feb 20 07:59:39 crc kubenswrapper[4492]: I0220 07:59:39.311076 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 07:59:39 crc kubenswrapper[4492]: I0220 07:59:39.311637 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 07:59:41 crc kubenswrapper[4492]: E0220 07:59:41.285917 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06e54b5b_dac1_4bc4_88b8_aa3558f84bf0.slice/crio-0483c58f2cf2595bd1b940d05bca54a902377c5bb999f91465afd29490cae026.scope\": RecentStats: unable to find data in memory cache]" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.168518 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pvjgb"] Feb 20 07:59:43 crc kubenswrapper[4492]: E0220 07:59:43.169628 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="registry-server" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.169672 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="registry-server" Feb 20 07:59:43 crc kubenswrapper[4492]: E0220 07:59:43.169689 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="extract-content" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.169695 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="extract-content" Feb 20 07:59:43 crc kubenswrapper[4492]: E0220 07:59:43.169737 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="registry-server" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.169743 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="registry-server" Feb 20 07:59:43 crc kubenswrapper[4492]: E0220 07:59:43.169761 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="extract-utilities" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.169768 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="extract-utilities" Feb 20 07:59:43 crc kubenswrapper[4492]: E0220 07:59:43.169782 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="extract-content" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.169788 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="extract-content" Feb 20 07:59:43 crc kubenswrapper[4492]: E0220 07:59:43.169802 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="extract-utilities" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.169808 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="extract-utilities" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.170030 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e54b5b-dac1-4bc4-88b8-aa3558f84bf0" containerName="registry-server" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.170070 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1bddb70-b394-482f-8993-62c877f8b055" containerName="registry-server" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.171570 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.180093 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pvjgb"] Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.374911 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-utilities\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.375011 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5rbl\" (UniqueName: \"kubernetes.io/projected/6282df32-3ee4-4cef-8275-5f939661571c-kube-api-access-z5rbl\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.375287 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-catalog-content\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.476331 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-catalog-content\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.476399 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-utilities\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.476439 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5rbl\" (UniqueName: \"kubernetes.io/projected/6282df32-3ee4-4cef-8275-5f939661571c-kube-api-access-z5rbl\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.476839 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-catalog-content\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.476886 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-utilities\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.494240 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5rbl\" (UniqueName: \"kubernetes.io/projected/6282df32-3ee4-4cef-8275-5f939661571c-kube-api-access-z5rbl\") pod \"redhat-operators-pvjgb\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.497919 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:43 crc kubenswrapper[4492]: I0220 07:59:43.923779 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pvjgb"] Feb 20 07:59:44 crc kubenswrapper[4492]: I0220 07:59:44.022443 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjgb" event={"ID":"6282df32-3ee4-4cef-8275-5f939661571c","Type":"ContainerStarted","Data":"84d935a03bb58115efafc92732c79dd52072e6ffdaee77bb17407ef3c7ef980a"} Feb 20 07:59:45 crc kubenswrapper[4492]: I0220 07:59:45.030990 4492 generic.go:334] "Generic (PLEG): container finished" podID="6282df32-3ee4-4cef-8275-5f939661571c" containerID="01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d" exitCode=0 Feb 20 07:59:45 crc kubenswrapper[4492]: I0220 07:59:45.031054 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjgb" event={"ID":"6282df32-3ee4-4cef-8275-5f939661571c","Type":"ContainerDied","Data":"01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d"} Feb 20 07:59:46 crc kubenswrapper[4492]: I0220 07:59:46.042398 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjgb" event={"ID":"6282df32-3ee4-4cef-8275-5f939661571c","Type":"ContainerStarted","Data":"74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6"} Feb 20 07:59:49 crc kubenswrapper[4492]: I0220 07:59:49.068081 4492 generic.go:334] "Generic (PLEG): container finished" podID="6282df32-3ee4-4cef-8275-5f939661571c" containerID="74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6" exitCode=0 Feb 20 07:59:49 crc kubenswrapper[4492]: I0220 07:59:49.068269 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjgb" event={"ID":"6282df32-3ee4-4cef-8275-5f939661571c","Type":"ContainerDied","Data":"74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6"} Feb 20 07:59:50 crc kubenswrapper[4492]: I0220 07:59:50.098404 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjgb" event={"ID":"6282df32-3ee4-4cef-8275-5f939661571c","Type":"ContainerStarted","Data":"bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc"} Feb 20 07:59:50 crc kubenswrapper[4492]: I0220 07:59:50.121347 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pvjgb" podStartSLOduration=2.611248481 podStartE2EDuration="7.121330887s" podCreationTimestamp="2026-02-20 07:59:43 +0000 UTC" firstStartedPulling="2026-02-20 07:59:45.032612167 +0000 UTC m=+4741.803901144" lastFinishedPulling="2026-02-20 07:59:49.542694571 +0000 UTC m=+4746.313983550" observedRunningTime="2026-02-20 07:59:50.115099596 +0000 UTC m=+4746.886388574" watchObservedRunningTime="2026-02-20 07:59:50.121330887 +0000 UTC m=+4746.892619864" Feb 20 07:59:53 crc kubenswrapper[4492]: I0220 07:59:53.498802 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:53 crc kubenswrapper[4492]: I0220 07:59:53.499288 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 07:59:54 crc kubenswrapper[4492]: I0220 07:59:54.538924 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pvjgb" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="registry-server" probeResult="failure" output=< Feb 20 07:59:54 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 07:59:54 crc kubenswrapper[4492]: > Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.301254 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh"] Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.303129 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.316862 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.316867 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.320185 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh"] Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.328368 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2c7017-b764-452f-9783-c8ab0b90fbdc-config-volume\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.328507 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2c7017-b764-452f-9783-c8ab0b90fbdc-secret-volume\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.328838 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvx8n\" (UniqueName: \"kubernetes.io/projected/3b2c7017-b764-452f-9783-c8ab0b90fbdc-kube-api-access-lvx8n\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.430792 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2c7017-b764-452f-9783-c8ab0b90fbdc-secret-volume\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.430858 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvx8n\" (UniqueName: \"kubernetes.io/projected/3b2c7017-b764-452f-9783-c8ab0b90fbdc-kube-api-access-lvx8n\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.430944 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2c7017-b764-452f-9783-c8ab0b90fbdc-config-volume\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.431880 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2c7017-b764-452f-9783-c8ab0b90fbdc-config-volume\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.441014 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2c7017-b764-452f-9783-c8ab0b90fbdc-secret-volume\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.448075 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvx8n\" (UniqueName: \"kubernetes.io/projected/3b2c7017-b764-452f-9783-c8ab0b90fbdc-kube-api-access-lvx8n\") pod \"collect-profiles-29526240-rjlvh\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:00 crc kubenswrapper[4492]: I0220 08:00:00.629131 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:01 crc kubenswrapper[4492]: I0220 08:00:01.172064 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh"] Feb 20 08:00:01 crc kubenswrapper[4492]: I0220 08:00:01.186158 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" event={"ID":"3b2c7017-b764-452f-9783-c8ab0b90fbdc","Type":"ContainerStarted","Data":"5a7edef371818259685cbffb9d12e8415f0d20a8868547ab6d0129500e30540f"} Feb 20 08:00:01 crc kubenswrapper[4492]: E0220 08:00:01.928401 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b2c7017_b764_452f_9783_c8ab0b90fbdc.slice/crio-c98c1a369d9eea98b825fbf139a60866c5cd77fc1de1d3eb8d469c8d58438d93.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b2c7017_b764_452f_9783_c8ab0b90fbdc.slice/crio-conmon-c98c1a369d9eea98b825fbf139a60866c5cd77fc1de1d3eb8d469c8d58438d93.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:00:02 crc kubenswrapper[4492]: I0220 08:00:02.197678 4492 generic.go:334] "Generic (PLEG): container finished" podID="3b2c7017-b764-452f-9783-c8ab0b90fbdc" containerID="c98c1a369d9eea98b825fbf139a60866c5cd77fc1de1d3eb8d469c8d58438d93" exitCode=0 Feb 20 08:00:02 crc kubenswrapper[4492]: I0220 08:00:02.197750 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" event={"ID":"3b2c7017-b764-452f-9783-c8ab0b90fbdc","Type":"ContainerDied","Data":"c98c1a369d9eea98b825fbf139a60866c5cd77fc1de1d3eb8d469c8d58438d93"} Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.650883 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.704683 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2c7017-b764-452f-9783-c8ab0b90fbdc-config-volume\") pod \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.704852 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2c7017-b764-452f-9783-c8ab0b90fbdc-secret-volume\") pod \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.704933 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvx8n\" (UniqueName: \"kubernetes.io/projected/3b2c7017-b764-452f-9783-c8ab0b90fbdc-kube-api-access-lvx8n\") pod \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\" (UID: \"3b2c7017-b764-452f-9783-c8ab0b90fbdc\") " Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.705726 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b2c7017-b764-452f-9783-c8ab0b90fbdc-config-volume" (OuterVolumeSpecName: "config-volume") pod "3b2c7017-b764-452f-9783-c8ab0b90fbdc" (UID: "3b2c7017-b764-452f-9783-c8ab0b90fbdc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.707823 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b2c7017-b764-452f-9783-c8ab0b90fbdc-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.715146 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b2c7017-b764-452f-9783-c8ab0b90fbdc-kube-api-access-lvx8n" (OuterVolumeSpecName: "kube-api-access-lvx8n") pod "3b2c7017-b764-452f-9783-c8ab0b90fbdc" (UID: "3b2c7017-b764-452f-9783-c8ab0b90fbdc"). InnerVolumeSpecName "kube-api-access-lvx8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.715616 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b2c7017-b764-452f-9783-c8ab0b90fbdc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3b2c7017-b764-452f-9783-c8ab0b90fbdc" (UID: "3b2c7017-b764-452f-9783-c8ab0b90fbdc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.809886 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvx8n\" (UniqueName: \"kubernetes.io/projected/3b2c7017-b764-452f-9783-c8ab0b90fbdc-kube-api-access-lvx8n\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:03 crc kubenswrapper[4492]: I0220 08:00:03.810044 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b2c7017-b764-452f-9783-c8ab0b90fbdc-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.212325 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" event={"ID":"3b2c7017-b764-452f-9783-c8ab0b90fbdc","Type":"ContainerDied","Data":"5a7edef371818259685cbffb9d12e8415f0d20a8868547ab6d0129500e30540f"} Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.212557 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a7edef371818259685cbffb9d12e8415f0d20a8868547ab6d0129500e30540f" Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.212367 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh" Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.555761 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pvjgb" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="registry-server" probeResult="failure" output=< Feb 20 08:00:04 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:00:04 crc kubenswrapper[4492]: > Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.721746 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv"] Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.728534 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526195-h6glv"] Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.976874 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wvh4d"] Feb 20 08:00:04 crc kubenswrapper[4492]: E0220 08:00:04.977263 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b2c7017-b764-452f-9783-c8ab0b90fbdc" containerName="collect-profiles" Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.977282 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b2c7017-b764-452f-9783-c8ab0b90fbdc" containerName="collect-profiles" Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.977438 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b2c7017-b764-452f-9783-c8ab0b90fbdc" containerName="collect-profiles" Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.982339 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:04 crc kubenswrapper[4492]: I0220 08:00:04.988733 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvh4d"] Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.030781 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-utilities\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.030948 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l94s\" (UniqueName: \"kubernetes.io/projected/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-kube-api-access-2l94s\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.031078 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-catalog-content\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.132318 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-utilities\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.132408 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l94s\" (UniqueName: \"kubernetes.io/projected/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-kube-api-access-2l94s\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.132485 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-catalog-content\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.132750 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-utilities\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.132997 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-catalog-content\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.148614 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l94s\" (UniqueName: \"kubernetes.io/projected/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-kube-api-access-2l94s\") pod \"redhat-marketplace-wvh4d\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.296759 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.566559 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8" path="/var/lib/kubelet/pods/be7a7d7e-da7c-49eb-ac8d-d8bffc3107c8/volumes" Feb 20 08:00:05 crc kubenswrapper[4492]: I0220 08:00:05.721267 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvh4d"] Feb 20 08:00:06 crc kubenswrapper[4492]: I0220 08:00:06.241963 4492 generic.go:334] "Generic (PLEG): container finished" podID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerID="adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446" exitCode=0 Feb 20 08:00:06 crc kubenswrapper[4492]: I0220 08:00:06.242109 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvh4d" event={"ID":"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff","Type":"ContainerDied","Data":"adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446"} Feb 20 08:00:06 crc kubenswrapper[4492]: I0220 08:00:06.242260 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvh4d" event={"ID":"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff","Type":"ContainerStarted","Data":"396fb25dea72d7fa35a2bfa2c9b3827f5ffb3c2708143132a2295d4e0dc53aa6"} Feb 20 08:00:08 crc kubenswrapper[4492]: I0220 08:00:08.259745 4492 generic.go:334] "Generic (PLEG): container finished" podID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerID="39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791" exitCode=0 Feb 20 08:00:08 crc kubenswrapper[4492]: I0220 08:00:08.259820 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvh4d" event={"ID":"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff","Type":"ContainerDied","Data":"39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791"} Feb 20 08:00:08 crc kubenswrapper[4492]: I0220 08:00:08.853944 4492 scope.go:117] "RemoveContainer" containerID="70214d122c4fe249cd6cbaf8e35355228ebdc9ad69ebe5eb885d5f0ea56382d7" Feb 20 08:00:09 crc kubenswrapper[4492]: I0220 08:00:09.279835 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvh4d" event={"ID":"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff","Type":"ContainerStarted","Data":"4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6"} Feb 20 08:00:09 crc kubenswrapper[4492]: I0220 08:00:09.311872 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:00:09 crc kubenswrapper[4492]: I0220 08:00:09.312134 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:00:09 crc kubenswrapper[4492]: I0220 08:00:09.312184 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:00:09 crc kubenswrapper[4492]: I0220 08:00:09.313505 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"046d2f99be3935b7577536a5396fd26213ad42cda70983181ddfca7d9b35ec0b"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:00:09 crc kubenswrapper[4492]: I0220 08:00:09.313735 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://046d2f99be3935b7577536a5396fd26213ad42cda70983181ddfca7d9b35ec0b" gracePeriod=600 Feb 20 08:00:09 crc kubenswrapper[4492]: I0220 08:00:09.313868 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wvh4d" podStartSLOduration=2.749862887 podStartE2EDuration="5.313845263s" podCreationTimestamp="2026-02-20 08:00:04 +0000 UTC" firstStartedPulling="2026-02-20 08:00:06.244914133 +0000 UTC m=+4763.016203110" lastFinishedPulling="2026-02-20 08:00:08.808896508 +0000 UTC m=+4765.580185486" observedRunningTime="2026-02-20 08:00:09.303766777 +0000 UTC m=+4766.075055756" watchObservedRunningTime="2026-02-20 08:00:09.313845263 +0000 UTC m=+4766.085134242" Feb 20 08:00:10 crc kubenswrapper[4492]: I0220 08:00:10.294062 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="046d2f99be3935b7577536a5396fd26213ad42cda70983181ddfca7d9b35ec0b" exitCode=0 Feb 20 08:00:10 crc kubenswrapper[4492]: I0220 08:00:10.294145 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"046d2f99be3935b7577536a5396fd26213ad42cda70983181ddfca7d9b35ec0b"} Feb 20 08:00:10 crc kubenswrapper[4492]: I0220 08:00:10.294696 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26"} Feb 20 08:00:10 crc kubenswrapper[4492]: I0220 08:00:10.294731 4492 scope.go:117] "RemoveContainer" containerID="12a91942a9c982f413dc9b75e534b157b71245f414f681cc35e2449ed9b21dda" Feb 20 08:00:13 crc kubenswrapper[4492]: I0220 08:00:13.536893 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 08:00:13 crc kubenswrapper[4492]: I0220 08:00:13.575239 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 08:00:14 crc kubenswrapper[4492]: I0220 08:00:14.371611 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pvjgb"] Feb 20 08:00:15 crc kubenswrapper[4492]: I0220 08:00:15.297406 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:15 crc kubenswrapper[4492]: I0220 08:00:15.297848 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:15 crc kubenswrapper[4492]: I0220 08:00:15.340305 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:15 crc kubenswrapper[4492]: I0220 08:00:15.369251 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pvjgb" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="registry-server" containerID="cri-o://bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc" gracePeriod=2 Feb 20 08:00:15 crc kubenswrapper[4492]: I0220 08:00:15.414441 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:15 crc kubenswrapper[4492]: I0220 08:00:15.981200 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.166403 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-utilities\") pod \"6282df32-3ee4-4cef-8275-5f939661571c\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.166456 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-catalog-content\") pod \"6282df32-3ee4-4cef-8275-5f939661571c\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.166623 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5rbl\" (UniqueName: \"kubernetes.io/projected/6282df32-3ee4-4cef-8275-5f939661571c-kube-api-access-z5rbl\") pod \"6282df32-3ee4-4cef-8275-5f939661571c\" (UID: \"6282df32-3ee4-4cef-8275-5f939661571c\") " Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.168490 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-utilities" (OuterVolumeSpecName: "utilities") pod "6282df32-3ee4-4cef-8275-5f939661571c" (UID: "6282df32-3ee4-4cef-8275-5f939661571c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.192036 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6282df32-3ee4-4cef-8275-5f939661571c-kube-api-access-z5rbl" (OuterVolumeSpecName: "kube-api-access-z5rbl") pod "6282df32-3ee4-4cef-8275-5f939661571c" (UID: "6282df32-3ee4-4cef-8275-5f939661571c"). InnerVolumeSpecName "kube-api-access-z5rbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.300967 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5rbl\" (UniqueName: \"kubernetes.io/projected/6282df32-3ee4-4cef-8275-5f939661571c-kube-api-access-z5rbl\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.301694 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.355938 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6282df32-3ee4-4cef-8275-5f939661571c" (UID: "6282df32-3ee4-4cef-8275-5f939661571c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.403819 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6282df32-3ee4-4cef-8275-5f939661571c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.403956 4492 generic.go:334] "Generic (PLEG): container finished" podID="6282df32-3ee4-4cef-8275-5f939661571c" containerID="bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc" exitCode=0 Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.404619 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pvjgb" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.405636 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjgb" event={"ID":"6282df32-3ee4-4cef-8275-5f939661571c","Type":"ContainerDied","Data":"bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc"} Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.405691 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pvjgb" event={"ID":"6282df32-3ee4-4cef-8275-5f939661571c","Type":"ContainerDied","Data":"84d935a03bb58115efafc92732c79dd52072e6ffdaee77bb17407ef3c7ef980a"} Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.405712 4492 scope.go:117] "RemoveContainer" containerID="bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.437331 4492 scope.go:117] "RemoveContainer" containerID="74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.454298 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pvjgb"] Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.461681 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pvjgb"] Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.497602 4492 scope.go:117] "RemoveContainer" containerID="01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.538544 4492 scope.go:117] "RemoveContainer" containerID="bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc" Feb 20 08:00:16 crc kubenswrapper[4492]: E0220 08:00:16.540892 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc\": container with ID starting with bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc not found: ID does not exist" containerID="bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.540948 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc"} err="failed to get container status \"bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc\": rpc error: code = NotFound desc = could not find container \"bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc\": container with ID starting with bcc8b494159168916f2fcaa804942dcd75e541f4670ab18fb3a45a1260eb75fc not found: ID does not exist" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.541005 4492 scope.go:117] "RemoveContainer" containerID="74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6" Feb 20 08:00:16 crc kubenswrapper[4492]: E0220 08:00:16.541363 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6\": container with ID starting with 74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6 not found: ID does not exist" containerID="74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.541390 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6"} err="failed to get container status \"74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6\": rpc error: code = NotFound desc = could not find container \"74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6\": container with ID starting with 74be227ce9949f09d114b25a1ea8472fab125379cb246ec0da3b46e51ce968a6 not found: ID does not exist" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.541408 4492 scope.go:117] "RemoveContainer" containerID="01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d" Feb 20 08:00:16 crc kubenswrapper[4492]: E0220 08:00:16.541676 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d\": container with ID starting with 01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d not found: ID does not exist" containerID="01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d" Feb 20 08:00:16 crc kubenswrapper[4492]: I0220 08:00:16.541698 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d"} err="failed to get container status \"01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d\": rpc error: code = NotFound desc = could not find container \"01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d\": container with ID starting with 01765b69327583b619acbb19f316a312cdbf1d51881a4e5ad3751ca42257c66d not found: ID does not exist" Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.375364 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvh4d"] Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.421764 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wvh4d" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="registry-server" containerID="cri-o://4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6" gracePeriod=2 Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.571891 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6282df32-3ee4-4cef-8275-5f939661571c" path="/var/lib/kubelet/pods/6282df32-3ee4-4cef-8275-5f939661571c/volumes" Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.893517 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.957826 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-catalog-content\") pod \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.958150 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-utilities\") pod \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.958468 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2l94s\" (UniqueName: \"kubernetes.io/projected/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-kube-api-access-2l94s\") pod \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\" (UID: \"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff\") " Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.959769 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-utilities" (OuterVolumeSpecName: "utilities") pod "fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" (UID: "fafb07ee-b29e-47e7-aed2-b9abc0acd6ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.965539 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-kube-api-access-2l94s" (OuterVolumeSpecName: "kube-api-access-2l94s") pod "fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" (UID: "fafb07ee-b29e-47e7-aed2-b9abc0acd6ff"). InnerVolumeSpecName "kube-api-access-2l94s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:00:17 crc kubenswrapper[4492]: I0220 08:00:17.985047 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" (UID: "fafb07ee-b29e-47e7-aed2-b9abc0acd6ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.062286 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.062324 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.062334 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2l94s\" (UniqueName: \"kubernetes.io/projected/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff-kube-api-access-2l94s\") on node \"crc\" DevicePath \"\"" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.432274 4492 generic.go:334] "Generic (PLEG): container finished" podID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerID="4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6" exitCode=0 Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.432328 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvh4d" event={"ID":"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff","Type":"ContainerDied","Data":"4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6"} Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.432384 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvh4d" event={"ID":"fafb07ee-b29e-47e7-aed2-b9abc0acd6ff","Type":"ContainerDied","Data":"396fb25dea72d7fa35a2bfa2c9b3827f5ffb3c2708143132a2295d4e0dc53aa6"} Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.432411 4492 scope.go:117] "RemoveContainer" containerID="4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.432434 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvh4d" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.468963 4492 scope.go:117] "RemoveContainer" containerID="39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.471913 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvh4d"] Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.485013 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvh4d"] Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.489459 4492 scope.go:117] "RemoveContainer" containerID="adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.521695 4492 scope.go:117] "RemoveContainer" containerID="4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6" Feb 20 08:00:18 crc kubenswrapper[4492]: E0220 08:00:18.522164 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6\": container with ID starting with 4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6 not found: ID does not exist" containerID="4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.522225 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6"} err="failed to get container status \"4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6\": rpc error: code = NotFound desc = could not find container \"4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6\": container with ID starting with 4f7a864820a4649829786951839749f8578cab38370866756a4ddc5959dc32d6 not found: ID does not exist" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.522264 4492 scope.go:117] "RemoveContainer" containerID="39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791" Feb 20 08:00:18 crc kubenswrapper[4492]: E0220 08:00:18.522730 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791\": container with ID starting with 39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791 not found: ID does not exist" containerID="39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.522763 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791"} err="failed to get container status \"39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791\": rpc error: code = NotFound desc = could not find container \"39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791\": container with ID starting with 39bb3984c459860d0a4270af420b92118bbb0247b7997811dc9b595a0089d791 not found: ID does not exist" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.522798 4492 scope.go:117] "RemoveContainer" containerID="adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446" Feb 20 08:00:18 crc kubenswrapper[4492]: E0220 08:00:18.523125 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446\": container with ID starting with adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446 not found: ID does not exist" containerID="adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446" Feb 20 08:00:18 crc kubenswrapper[4492]: I0220 08:00:18.523178 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446"} err="failed to get container status \"adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446\": rpc error: code = NotFound desc = could not find container \"adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446\": container with ID starting with adca96088117732b8847483d0d1da48b6eabe5c418657822d4eb65b1c4b48446 not found: ID does not exist" Feb 20 08:00:19 crc kubenswrapper[4492]: I0220 08:00:19.576644 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" path="/var/lib/kubelet/pods/fafb07ee-b29e-47e7-aed2-b9abc0acd6ff/volumes" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.152640 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29526241-fgz9q"] Feb 20 08:01:00 crc kubenswrapper[4492]: E0220 08:01:00.153684 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="extract-content" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.153701 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="extract-content" Feb 20 08:01:00 crc kubenswrapper[4492]: E0220 08:01:00.153724 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="extract-utilities" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.153729 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="extract-utilities" Feb 20 08:01:00 crc kubenswrapper[4492]: E0220 08:01:00.153740 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="registry-server" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.153746 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="registry-server" Feb 20 08:01:00 crc kubenswrapper[4492]: E0220 08:01:00.153754 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="registry-server" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.153759 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="registry-server" Feb 20 08:01:00 crc kubenswrapper[4492]: E0220 08:01:00.153774 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="extract-utilities" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.153779 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="extract-utilities" Feb 20 08:01:00 crc kubenswrapper[4492]: E0220 08:01:00.153802 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="extract-content" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.153807 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="extract-content" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.154008 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="6282df32-3ee4-4cef-8275-5f939661571c" containerName="registry-server" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.154035 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="fafb07ee-b29e-47e7-aed2-b9abc0acd6ff" containerName="registry-server" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.154715 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.166809 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526241-fgz9q"] Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.257130 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-config-data\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.257193 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlgtg\" (UniqueName: \"kubernetes.io/projected/c8aff97d-0465-4db5-933d-d67cfa494670-kube-api-access-mlgtg\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.257302 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-combined-ca-bundle\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.257325 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-fernet-keys\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.359807 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-combined-ca-bundle\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.359984 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-fernet-keys\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.360246 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-config-data\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.360376 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlgtg\" (UniqueName: \"kubernetes.io/projected/c8aff97d-0465-4db5-933d-d67cfa494670-kube-api-access-mlgtg\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.371562 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-fernet-keys\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.371725 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-combined-ca-bundle\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.373426 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-config-data\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.375356 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlgtg\" (UniqueName: \"kubernetes.io/projected/c8aff97d-0465-4db5-933d-d67cfa494670-kube-api-access-mlgtg\") pod \"keystone-cron-29526241-fgz9q\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.478210 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:00 crc kubenswrapper[4492]: I0220 08:01:00.985918 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526241-fgz9q"] Feb 20 08:01:01 crc kubenswrapper[4492]: I0220 08:01:01.838436 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526241-fgz9q" event={"ID":"c8aff97d-0465-4db5-933d-d67cfa494670","Type":"ContainerStarted","Data":"91b9413e4a6e317f655ebf0329a8f463a90217d1eac6d08c81edffd1f306641a"} Feb 20 08:01:01 crc kubenswrapper[4492]: I0220 08:01:01.839037 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526241-fgz9q" event={"ID":"c8aff97d-0465-4db5-933d-d67cfa494670","Type":"ContainerStarted","Data":"7dc8ce6867bbf3755b704cf57feea744c2bc9d50f607b22fa8c358ed9f8a92fc"} Feb 20 08:01:01 crc kubenswrapper[4492]: I0220 08:01:01.860740 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29526241-fgz9q" podStartSLOduration=1.860722944 podStartE2EDuration="1.860722944s" podCreationTimestamp="2026-02-20 08:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:01:01.857801965 +0000 UTC m=+4818.629090943" watchObservedRunningTime="2026-02-20 08:01:01.860722944 +0000 UTC m=+4818.632011922" Feb 20 08:01:03 crc kubenswrapper[4492]: E0220 08:01:03.337361 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8aff97d_0465_4db5_933d_d67cfa494670.slice/crio-91b9413e4a6e317f655ebf0329a8f463a90217d1eac6d08c81edffd1f306641a.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:01:03 crc kubenswrapper[4492]: I0220 08:01:03.859258 4492 generic.go:334] "Generic (PLEG): container finished" podID="c8aff97d-0465-4db5-933d-d67cfa494670" containerID="91b9413e4a6e317f655ebf0329a8f463a90217d1eac6d08c81edffd1f306641a" exitCode=0 Feb 20 08:01:03 crc kubenswrapper[4492]: I0220 08:01:03.859342 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526241-fgz9q" event={"ID":"c8aff97d-0465-4db5-933d-d67cfa494670","Type":"ContainerDied","Data":"91b9413e4a6e317f655ebf0329a8f463a90217d1eac6d08c81edffd1f306641a"} Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.155630 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.280321 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlgtg\" (UniqueName: \"kubernetes.io/projected/c8aff97d-0465-4db5-933d-d67cfa494670-kube-api-access-mlgtg\") pod \"c8aff97d-0465-4db5-933d-d67cfa494670\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.280575 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-combined-ca-bundle\") pod \"c8aff97d-0465-4db5-933d-d67cfa494670\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.280613 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-config-data\") pod \"c8aff97d-0465-4db5-933d-d67cfa494670\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.280798 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-fernet-keys\") pod \"c8aff97d-0465-4db5-933d-d67cfa494670\" (UID: \"c8aff97d-0465-4db5-933d-d67cfa494670\") " Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.287375 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c8aff97d-0465-4db5-933d-d67cfa494670" (UID: "c8aff97d-0465-4db5-933d-d67cfa494670"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.288734 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8aff97d-0465-4db5-933d-d67cfa494670-kube-api-access-mlgtg" (OuterVolumeSpecName: "kube-api-access-mlgtg") pod "c8aff97d-0465-4db5-933d-d67cfa494670" (UID: "c8aff97d-0465-4db5-933d-d67cfa494670"). InnerVolumeSpecName "kube-api-access-mlgtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.307787 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8aff97d-0465-4db5-933d-d67cfa494670" (UID: "c8aff97d-0465-4db5-933d-d67cfa494670"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.331022 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-config-data" (OuterVolumeSpecName: "config-data") pod "c8aff97d-0465-4db5-933d-d67cfa494670" (UID: "c8aff97d-0465-4db5-933d-d67cfa494670"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.385396 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.385430 4492 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.385455 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlgtg\" (UniqueName: \"kubernetes.io/projected/c8aff97d-0465-4db5-933d-d67cfa494670-kube-api-access-mlgtg\") on node \"crc\" DevicePath \"\"" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.385468 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8aff97d-0465-4db5-933d-d67cfa494670-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.875672 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526241-fgz9q" event={"ID":"c8aff97d-0465-4db5-933d-d67cfa494670","Type":"ContainerDied","Data":"7dc8ce6867bbf3755b704cf57feea744c2bc9d50f607b22fa8c358ed9f8a92fc"} Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.876027 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dc8ce6867bbf3755b704cf57feea744c2bc9d50f607b22fa8c358ed9f8a92fc" Feb 20 08:01:05 crc kubenswrapper[4492]: I0220 08:01:05.875736 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526241-fgz9q" Feb 20 08:02:09 crc kubenswrapper[4492]: I0220 08:02:09.311414 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:02:09 crc kubenswrapper[4492]: I0220 08:02:09.313306 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:02:39 crc kubenswrapper[4492]: I0220 08:02:39.311766 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:02:39 crc kubenswrapper[4492]: I0220 08:02:39.312518 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.310819 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.311148 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.311195 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.311943 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.311990 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" gracePeriod=600 Feb 20 08:03:09 crc kubenswrapper[4492]: E0220 08:03:09.428432 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.940218 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" exitCode=0 Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.940289 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26"} Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.940626 4492 scope.go:117] "RemoveContainer" containerID="046d2f99be3935b7577536a5396fd26213ad42cda70983181ddfca7d9b35ec0b" Feb 20 08:03:09 crc kubenswrapper[4492]: I0220 08:03:09.941272 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:03:09 crc kubenswrapper[4492]: E0220 08:03:09.941634 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:03:24 crc kubenswrapper[4492]: I0220 08:03:24.557206 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:03:24 crc kubenswrapper[4492]: E0220 08:03:24.558017 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:03:37 crc kubenswrapper[4492]: I0220 08:03:37.557664 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:03:37 crc kubenswrapper[4492]: E0220 08:03:37.558749 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:03:50 crc kubenswrapper[4492]: I0220 08:03:50.556703 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:03:50 crc kubenswrapper[4492]: E0220 08:03:50.557423 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:04:02 crc kubenswrapper[4492]: I0220 08:04:02.557508 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:04:02 crc kubenswrapper[4492]: E0220 08:04:02.558269 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:04:17 crc kubenswrapper[4492]: I0220 08:04:17.557440 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:04:17 crc kubenswrapper[4492]: E0220 08:04:17.558422 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:04:30 crc kubenswrapper[4492]: I0220 08:04:30.557168 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:04:30 crc kubenswrapper[4492]: E0220 08:04:30.558095 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:04:42 crc kubenswrapper[4492]: I0220 08:04:42.557126 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:04:42 crc kubenswrapper[4492]: E0220 08:04:42.558073 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:04:57 crc kubenswrapper[4492]: I0220 08:04:57.557697 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:04:57 crc kubenswrapper[4492]: E0220 08:04:57.558577 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:05:11 crc kubenswrapper[4492]: I0220 08:05:11.557378 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:05:11 crc kubenswrapper[4492]: E0220 08:05:11.558331 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:05:25 crc kubenswrapper[4492]: I0220 08:05:25.556849 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:05:25 crc kubenswrapper[4492]: E0220 08:05:25.567931 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:05:38 crc kubenswrapper[4492]: I0220 08:05:38.557167 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:05:38 crc kubenswrapper[4492]: E0220 08:05:38.558074 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:05:44 crc kubenswrapper[4492]: E0220 08:05:44.390112 4492 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.25:54746->192.168.26.25:44007: write tcp 192.168.26.25:54746->192.168.26.25:44007: write: broken pipe Feb 20 08:05:53 crc kubenswrapper[4492]: I0220 08:05:53.564985 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:05:53 crc kubenswrapper[4492]: E0220 08:05:53.565907 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:06:08 crc kubenswrapper[4492]: I0220 08:06:08.557842 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:06:08 crc kubenswrapper[4492]: E0220 08:06:08.558743 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:06:19 crc kubenswrapper[4492]: I0220 08:06:19.558625 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:06:19 crc kubenswrapper[4492]: E0220 08:06:19.559529 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:06:19 crc kubenswrapper[4492]: I0220 08:06:19.720905 4492 generic.go:334] "Generic (PLEG): container finished" podID="419ec593-81d8-44f8-9d19-5df9ebf183e1" containerID="01d4922c7c8ea5fe9cf1952e2a8cbb04bac6d0f58f7b3cf534717314f9c3f53f" exitCode=0 Feb 20 08:06:19 crc kubenswrapper[4492]: I0220 08:06:19.721519 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"419ec593-81d8-44f8-9d19-5df9ebf183e1","Type":"ContainerDied","Data":"01d4922c7c8ea5fe9cf1952e2a8cbb04bac6d0f58f7b3cf534717314f9c3f53f"} Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.372620 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.474824 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Feb 20 08:06:21 crc kubenswrapper[4492]: E0220 08:06:21.475592 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8aff97d-0465-4db5-933d-d67cfa494670" containerName="keystone-cron" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.475726 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8aff97d-0465-4db5-933d-d67cfa494670" containerName="keystone-cron" Feb 20 08:06:21 crc kubenswrapper[4492]: E0220 08:06:21.475789 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="419ec593-81d8-44f8-9d19-5df9ebf183e1" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.475860 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="419ec593-81d8-44f8-9d19-5df9ebf183e1" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.476289 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8aff97d-0465-4db5-933d-d67cfa494670" containerName="keystone-cron" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.476424 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="419ec593-81d8-44f8-9d19-5df9ebf183e1" containerName="tempest-tests-tempest-tests-runner" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.477277 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.490114 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.490294 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.493563 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.510906 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ssh-key\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511118 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config-secret\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511226 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511324 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-workdir\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511403 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-config-data\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511443 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm5ff\" (UniqueName: \"kubernetes.io/projected/419ec593-81d8-44f8-9d19-5df9ebf183e1-kube-api-access-xm5ff\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511489 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ca-certs\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511588 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.511657 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-temporary\") pod \"419ec593-81d8-44f8-9d19-5df9ebf183e1\" (UID: \"419ec593-81d8-44f8-9d19-5df9ebf183e1\") " Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.523797 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-config-data" (OuterVolumeSpecName: "config-data") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.525488 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.533034 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.533948 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/419ec593-81d8-44f8-9d19-5df9ebf183e1-kube-api-access-xm5ff" (OuterVolumeSpecName: "kube-api-access-xm5ff") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "kube-api-access-xm5ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.541440 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.549120 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.555054 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.555361 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.566786 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "419ec593-81d8-44f8-9d19-5df9ebf183e1" (UID: "419ec593-81d8-44f8-9d19-5df9ebf183e1"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.614197 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.614396 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.614611 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.614713 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.614802 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dpsg\" (UniqueName: \"kubernetes.io/projected/0d1fd559-3c53-44c1-aa9b-556b306a479b-kube-api-access-6dpsg\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.614900 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615011 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615130 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615322 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615485 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615552 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615605 4492 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615665 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615749 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm5ff\" (UniqueName: \"kubernetes.io/projected/419ec593-81d8-44f8-9d19-5df9ebf183e1-kube-api-access-xm5ff\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615801 4492 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/419ec593-81d8-44f8-9d19-5df9ebf183e1-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615854 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/419ec593-81d8-44f8-9d19-5df9ebf183e1-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.615906 4492 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/419ec593-81d8-44f8-9d19-5df9ebf183e1-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.641256 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.718022 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.718209 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.718320 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dpsg\" (UniqueName: \"kubernetes.io/projected/0d1fd559-3c53-44c1-aa9b-556b306a479b-kube-api-access-6dpsg\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.719142 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.719314 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.719405 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.719444 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.719505 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.721308 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.721836 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.721934 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.722008 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.722534 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.722943 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.723555 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.734858 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dpsg\" (UniqueName: \"kubernetes.io/projected/0d1fd559-3c53-44c1-aa9b-556b306a479b-kube-api-access-6dpsg\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.740852 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"419ec593-81d8-44f8-9d19-5df9ebf183e1","Type":"ContainerDied","Data":"a94ba04d5aabd2da6882f3d8a9211007bb2cc78aeb1dfab60e62ba84b39e265a"} Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.740889 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a94ba04d5aabd2da6882f3d8a9211007bb2cc78aeb1dfab60e62ba84b39e265a" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.740944 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Feb 20 08:06:21 crc kubenswrapper[4492]: I0220 08:06:21.814061 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 08:06:22 crc kubenswrapper[4492]: I0220 08:06:22.343197 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Feb 20 08:06:22 crc kubenswrapper[4492]: W0220 08:06:22.359605 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d1fd559_3c53_44c1_aa9b_556b306a479b.slice/crio-1afa41b1fe42bc6ab6c4d34bdd4030962a3db49ebd170cf869e7575aa8fca2a7 WatchSource:0}: Error finding container 1afa41b1fe42bc6ab6c4d34bdd4030962a3db49ebd170cf869e7575aa8fca2a7: Status 404 returned error can't find the container with id 1afa41b1fe42bc6ab6c4d34bdd4030962a3db49ebd170cf869e7575aa8fca2a7 Feb 20 08:06:22 crc kubenswrapper[4492]: I0220 08:06:22.751321 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"0d1fd559-3c53-44c1-aa9b-556b306a479b","Type":"ContainerStarted","Data":"1afa41b1fe42bc6ab6c4d34bdd4030962a3db49ebd170cf869e7575aa8fca2a7"} Feb 20 08:06:24 crc kubenswrapper[4492]: I0220 08:06:24.775839 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"0d1fd559-3c53-44c1-aa9b-556b306a479b","Type":"ContainerStarted","Data":"33b4912328c192d6e2377bb42694cffe83780ea6af9e48234e0fc59923cd57db"} Feb 20 08:06:24 crc kubenswrapper[4492]: I0220 08:06:24.799318 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" podStartSLOduration=3.799303713 podStartE2EDuration="3.799303713s" podCreationTimestamp="2026-02-20 08:06:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:06:24.793196566 +0000 UTC m=+5141.564485544" watchObservedRunningTime="2026-02-20 08:06:24.799303713 +0000 UTC m=+5141.570592691" Feb 20 08:06:33 crc kubenswrapper[4492]: I0220 08:06:33.570584 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:06:33 crc kubenswrapper[4492]: E0220 08:06:33.571937 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:06:47 crc kubenswrapper[4492]: I0220 08:06:47.561787 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:06:47 crc kubenswrapper[4492]: E0220 08:06:47.562488 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:06:58 crc kubenswrapper[4492]: I0220 08:06:58.557884 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:06:58 crc kubenswrapper[4492]: E0220 08:06:58.558618 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.557118 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:07:10 crc kubenswrapper[4492]: E0220 08:07:10.557811 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.571130 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-68fcdb65f7-pbqcq"] Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.572450 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.598818 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68fcdb65f7-pbqcq"] Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.758314 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-httpd-config\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.758380 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-config\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.758427 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-combined-ca-bundle\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.758491 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w68zl\" (UniqueName: \"kubernetes.io/projected/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-kube-api-access-w68zl\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.758671 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-internal-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.758764 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-public-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.759072 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-ovndb-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.862434 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-ovndb-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.862996 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-httpd-config\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.863124 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-config\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.863291 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-combined-ca-bundle\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.863439 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w68zl\" (UniqueName: \"kubernetes.io/projected/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-kube-api-access-w68zl\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.864731 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-internal-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.864842 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-public-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.876883 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-internal-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.878294 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-httpd-config\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.878602 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-combined-ca-bundle\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.879411 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-ovndb-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.880085 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-config\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.882767 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-public-tls-certs\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.887308 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w68zl\" (UniqueName: \"kubernetes.io/projected/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-kube-api-access-w68zl\") pod \"neutron-68fcdb65f7-pbqcq\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:10 crc kubenswrapper[4492]: I0220 08:07:10.890550 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:11 crc kubenswrapper[4492]: I0220 08:07:11.729735 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-68fcdb65f7-pbqcq"] Feb 20 08:07:12 crc kubenswrapper[4492]: I0220 08:07:12.260860 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68fcdb65f7-pbqcq" event={"ID":"3f540de0-30e4-4b8b-b8de-628f4d46aa4c","Type":"ContainerStarted","Data":"55fc28fed46a899b668032714898309dd768815e5a9f5eb431a67a6f1574ce03"} Feb 20 08:07:12 crc kubenswrapper[4492]: I0220 08:07:12.261227 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68fcdb65f7-pbqcq" event={"ID":"3f540de0-30e4-4b8b-b8de-628f4d46aa4c","Type":"ContainerStarted","Data":"3f56ae764876ab87e13df3f6dd4387d92194dea3e5856e1f77114c7efee5a7ec"} Feb 20 08:07:12 crc kubenswrapper[4492]: I0220 08:07:12.261511 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:12 crc kubenswrapper[4492]: I0220 08:07:12.261544 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68fcdb65f7-pbqcq" event={"ID":"3f540de0-30e4-4b8b-b8de-628f4d46aa4c","Type":"ContainerStarted","Data":"071741118bfdd7557b22826e3e2d2ccf495bc3fb30a14cfa81a16a7da77912f4"} Feb 20 08:07:12 crc kubenswrapper[4492]: I0220 08:07:12.288222 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-68fcdb65f7-pbqcq" podStartSLOduration=2.28820674 podStartE2EDuration="2.28820674s" podCreationTimestamp="2026-02-20 08:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:07:12.281062077 +0000 UTC m=+5189.052351054" watchObservedRunningTime="2026-02-20 08:07:12.28820674 +0000 UTC m=+5189.059495718" Feb 20 08:07:24 crc kubenswrapper[4492]: I0220 08:07:24.558002 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:07:24 crc kubenswrapper[4492]: E0220 08:07:24.559382 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:07:35 crc kubenswrapper[4492]: I0220 08:07:35.557132 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:07:35 crc kubenswrapper[4492]: E0220 08:07:35.558173 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:07:40 crc kubenswrapper[4492]: I0220 08:07:40.906322 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:07:40 crc kubenswrapper[4492]: I0220 08:07:40.963819 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65749686ff-sv9sq"] Feb 20 08:07:40 crc kubenswrapper[4492]: I0220 08:07:40.964039 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65749686ff-sv9sq" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-api" containerID="cri-o://022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1" gracePeriod=30 Feb 20 08:07:40 crc kubenswrapper[4492]: I0220 08:07:40.964261 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65749686ff-sv9sq" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-httpd" containerID="cri-o://b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b" gracePeriod=30 Feb 20 08:07:41 crc kubenswrapper[4492]: I0220 08:07:41.520539 4492 generic.go:334] "Generic (PLEG): container finished" podID="495833af-a575-4097-9033-e968a74f3a18" containerID="b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b" exitCode=0 Feb 20 08:07:41 crc kubenswrapper[4492]: I0220 08:07:41.520589 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65749686ff-sv9sq" event={"ID":"495833af-a575-4097-9033-e968a74f3a18","Type":"ContainerDied","Data":"b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b"} Feb 20 08:07:43 crc kubenswrapper[4492]: I0220 08:07:43.929481 4492 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-65749686ff-sv9sq" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.163:9696/\": dial tcp 10.217.0.163:9696: connect: connection refused" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.210282 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65749686ff-sv9sq" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.350800 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-combined-ca-bundle\") pod \"495833af-a575-4097-9033-e968a74f3a18\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.350892 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-ovndb-tls-certs\") pod \"495833af-a575-4097-9033-e968a74f3a18\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.350918 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99zcp\" (UniqueName: \"kubernetes.io/projected/495833af-a575-4097-9033-e968a74f3a18-kube-api-access-99zcp\") pod \"495833af-a575-4097-9033-e968a74f3a18\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.350938 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-httpd-config\") pod \"495833af-a575-4097-9033-e968a74f3a18\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.350955 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-public-tls-certs\") pod \"495833af-a575-4097-9033-e968a74f3a18\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.350980 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-config\") pod \"495833af-a575-4097-9033-e968a74f3a18\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.351057 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-internal-tls-certs\") pod \"495833af-a575-4097-9033-e968a74f3a18\" (UID: \"495833af-a575-4097-9033-e968a74f3a18\") " Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.358815 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "495833af-a575-4097-9033-e968a74f3a18" (UID: "495833af-a575-4097-9033-e968a74f3a18"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.359056 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495833af-a575-4097-9033-e968a74f3a18-kube-api-access-99zcp" (OuterVolumeSpecName: "kube-api-access-99zcp") pod "495833af-a575-4097-9033-e968a74f3a18" (UID: "495833af-a575-4097-9033-e968a74f3a18"). InnerVolumeSpecName "kube-api-access-99zcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.400964 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "495833af-a575-4097-9033-e968a74f3a18" (UID: "495833af-a575-4097-9033-e968a74f3a18"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.406209 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "495833af-a575-4097-9033-e968a74f3a18" (UID: "495833af-a575-4097-9033-e968a74f3a18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.412767 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-config" (OuterVolumeSpecName: "config") pod "495833af-a575-4097-9033-e968a74f3a18" (UID: "495833af-a575-4097-9033-e968a74f3a18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.420201 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "495833af-a575-4097-9033-e968a74f3a18" (UID: "495833af-a575-4097-9033-e968a74f3a18"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.429168 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "495833af-a575-4097-9033-e968a74f3a18" (UID: "495833af-a575-4097-9033-e968a74f3a18"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.453751 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.453786 4492 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.453797 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99zcp\" (UniqueName: \"kubernetes.io/projected/495833af-a575-4097-9033-e968a74f3a18-kube-api-access-99zcp\") on node \"crc\" DevicePath \"\"" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.453808 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.453817 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.453826 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.453836 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/495833af-a575-4097-9033-e968a74f3a18-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.560677 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:07:49 crc kubenswrapper[4492]: E0220 08:07:49.561164 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.592407 4492 generic.go:334] "Generic (PLEG): container finished" podID="495833af-a575-4097-9033-e968a74f3a18" containerID="022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1" exitCode=0 Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.592491 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65749686ff-sv9sq" event={"ID":"495833af-a575-4097-9033-e968a74f3a18","Type":"ContainerDied","Data":"022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1"} Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.592529 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65749686ff-sv9sq" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.592805 4492 scope.go:117] "RemoveContainer" containerID="b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.592695 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65749686ff-sv9sq" event={"ID":"495833af-a575-4097-9033-e968a74f3a18","Type":"ContainerDied","Data":"f8851eb0435e46350e5644f98d9320468afe894129862ee6e917b2c442bb4669"} Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.615063 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65749686ff-sv9sq"] Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.624459 4492 scope.go:117] "RemoveContainer" containerID="022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.624795 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-65749686ff-sv9sq"] Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.652676 4492 scope.go:117] "RemoveContainer" containerID="b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b" Feb 20 08:07:49 crc kubenswrapper[4492]: E0220 08:07:49.653438 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b\": container with ID starting with b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b not found: ID does not exist" containerID="b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.653497 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b"} err="failed to get container status \"b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b\": rpc error: code = NotFound desc = could not find container \"b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b\": container with ID starting with b9dbd96951384a715e8cf38004e41a391d576abc1496273676d54efa01f2b04b not found: ID does not exist" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.653525 4492 scope.go:117] "RemoveContainer" containerID="022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1" Feb 20 08:07:49 crc kubenswrapper[4492]: E0220 08:07:49.654286 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1\": container with ID starting with 022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1 not found: ID does not exist" containerID="022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1" Feb 20 08:07:49 crc kubenswrapper[4492]: I0220 08:07:49.654332 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1"} err="failed to get container status \"022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1\": rpc error: code = NotFound desc = could not find container \"022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1\": container with ID starting with 022ccc98e59105e1793c123781061068560360f295a3742b279fd403bcba35e1 not found: ID does not exist" Feb 20 08:07:51 crc kubenswrapper[4492]: I0220 08:07:51.568703 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="495833af-a575-4097-9033-e968a74f3a18" path="/var/lib/kubelet/pods/495833af-a575-4097-9033-e968a74f3a18/volumes" Feb 20 08:08:04 crc kubenswrapper[4492]: I0220 08:08:04.556876 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:08:04 crc kubenswrapper[4492]: E0220 08:08:04.557894 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:08:16 crc kubenswrapper[4492]: I0220 08:08:16.556531 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:08:16 crc kubenswrapper[4492]: I0220 08:08:16.866515 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"53c62c1dc0f3353ec3654a5971e2430befc77fd708a235e5633ff86ec1759add"} Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.852606 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cjmvb"] Feb 20 08:09:23 crc kubenswrapper[4492]: E0220 08:09:23.854078 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-httpd" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.854095 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-httpd" Feb 20 08:09:23 crc kubenswrapper[4492]: E0220 08:09:23.854111 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-api" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.854117 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-api" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.854304 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-api" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.854318 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="495833af-a575-4097-9033-e968a74f3a18" containerName="neutron-httpd" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.856458 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.867941 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-utilities\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.868192 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj4df\" (UniqueName: \"kubernetes.io/projected/469df4a7-7e1b-434c-9605-f79bb68c196c-kube-api-access-bj4df\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.868249 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-catalog-content\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.877869 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjmvb"] Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.971227 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-utilities\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.971573 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj4df\" (UniqueName: \"kubernetes.io/projected/469df4a7-7e1b-434c-9605-f79bb68c196c-kube-api-access-bj4df\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.971709 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-catalog-content\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.971733 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-utilities\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.972064 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-catalog-content\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:23 crc kubenswrapper[4492]: I0220 08:09:23.990280 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj4df\" (UniqueName: \"kubernetes.io/projected/469df4a7-7e1b-434c-9605-f79bb68c196c-kube-api-access-bj4df\") pod \"certified-operators-cjmvb\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:24 crc kubenswrapper[4492]: I0220 08:09:24.175103 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:24 crc kubenswrapper[4492]: I0220 08:09:24.586987 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjmvb"] Feb 20 08:09:25 crc kubenswrapper[4492]: I0220 08:09:25.473357 4492 generic.go:334] "Generic (PLEG): container finished" podID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerID="f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c" exitCode=0 Feb 20 08:09:25 crc kubenswrapper[4492]: I0220 08:09:25.473458 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjmvb" event={"ID":"469df4a7-7e1b-434c-9605-f79bb68c196c","Type":"ContainerDied","Data":"f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c"} Feb 20 08:09:25 crc kubenswrapper[4492]: I0220 08:09:25.473710 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjmvb" event={"ID":"469df4a7-7e1b-434c-9605-f79bb68c196c","Type":"ContainerStarted","Data":"83177ab8b9d01d6b5860bf9f8a3e3672e7e84344de9fa09491e436bcacb586f1"} Feb 20 08:09:25 crc kubenswrapper[4492]: I0220 08:09:25.477752 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:09:26 crc kubenswrapper[4492]: I0220 08:09:26.486557 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjmvb" event={"ID":"469df4a7-7e1b-434c-9605-f79bb68c196c","Type":"ContainerStarted","Data":"52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178"} Feb 20 08:09:28 crc kubenswrapper[4492]: I0220 08:09:28.503021 4492 generic.go:334] "Generic (PLEG): container finished" podID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerID="52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178" exitCode=0 Feb 20 08:09:28 crc kubenswrapper[4492]: I0220 08:09:28.503100 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjmvb" event={"ID":"469df4a7-7e1b-434c-9605-f79bb68c196c","Type":"ContainerDied","Data":"52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178"} Feb 20 08:09:29 crc kubenswrapper[4492]: I0220 08:09:29.515853 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjmvb" event={"ID":"469df4a7-7e1b-434c-9605-f79bb68c196c","Type":"ContainerStarted","Data":"37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727"} Feb 20 08:09:29 crc kubenswrapper[4492]: I0220 08:09:29.542750 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cjmvb" podStartSLOduration=3.043191043 podStartE2EDuration="6.542732469s" podCreationTimestamp="2026-02-20 08:09:23 +0000 UTC" firstStartedPulling="2026-02-20 08:09:25.475812871 +0000 UTC m=+5322.247101849" lastFinishedPulling="2026-02-20 08:09:28.975354298 +0000 UTC m=+5325.746643275" observedRunningTime="2026-02-20 08:09:29.536057692 +0000 UTC m=+5326.307346660" watchObservedRunningTime="2026-02-20 08:09:29.542732469 +0000 UTC m=+5326.314021448" Feb 20 08:09:34 crc kubenswrapper[4492]: I0220 08:09:34.175274 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:34 crc kubenswrapper[4492]: I0220 08:09:34.175799 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:34 crc kubenswrapper[4492]: I0220 08:09:34.216853 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:34 crc kubenswrapper[4492]: I0220 08:09:34.591967 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:34 crc kubenswrapper[4492]: I0220 08:09:34.636362 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjmvb"] Feb 20 08:09:36 crc kubenswrapper[4492]: I0220 08:09:36.571303 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cjmvb" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="registry-server" containerID="cri-o://37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727" gracePeriod=2 Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.018348 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.133701 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-utilities\") pod \"469df4a7-7e1b-434c-9605-f79bb68c196c\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.134060 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-catalog-content\") pod \"469df4a7-7e1b-434c-9605-f79bb68c196c\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.134614 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-utilities" (OuterVolumeSpecName: "utilities") pod "469df4a7-7e1b-434c-9605-f79bb68c196c" (UID: "469df4a7-7e1b-434c-9605-f79bb68c196c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.134730 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj4df\" (UniqueName: \"kubernetes.io/projected/469df4a7-7e1b-434c-9605-f79bb68c196c-kube-api-access-bj4df\") pod \"469df4a7-7e1b-434c-9605-f79bb68c196c\" (UID: \"469df4a7-7e1b-434c-9605-f79bb68c196c\") " Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.136087 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.140234 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469df4a7-7e1b-434c-9605-f79bb68c196c-kube-api-access-bj4df" (OuterVolumeSpecName: "kube-api-access-bj4df") pod "469df4a7-7e1b-434c-9605-f79bb68c196c" (UID: "469df4a7-7e1b-434c-9605-f79bb68c196c"). InnerVolumeSpecName "kube-api-access-bj4df". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.176951 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "469df4a7-7e1b-434c-9605-f79bb68c196c" (UID: "469df4a7-7e1b-434c-9605-f79bb68c196c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.238214 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469df4a7-7e1b-434c-9605-f79bb68c196c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.238525 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj4df\" (UniqueName: \"kubernetes.io/projected/469df4a7-7e1b-434c-9605-f79bb68c196c-kube-api-access-bj4df\") on node \"crc\" DevicePath \"\"" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.590098 4492 generic.go:334] "Generic (PLEG): container finished" podID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerID="37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727" exitCode=0 Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.590162 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjmvb" event={"ID":"469df4a7-7e1b-434c-9605-f79bb68c196c","Type":"ContainerDied","Data":"37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727"} Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.590189 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjmvb" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.590208 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjmvb" event={"ID":"469df4a7-7e1b-434c-9605-f79bb68c196c","Type":"ContainerDied","Data":"83177ab8b9d01d6b5860bf9f8a3e3672e7e84344de9fa09491e436bcacb586f1"} Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.590241 4492 scope.go:117] "RemoveContainer" containerID="37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.620502 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjmvb"] Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.624557 4492 scope.go:117] "RemoveContainer" containerID="52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.627663 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cjmvb"] Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.647612 4492 scope.go:117] "RemoveContainer" containerID="f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.678816 4492 scope.go:117] "RemoveContainer" containerID="37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727" Feb 20 08:09:37 crc kubenswrapper[4492]: E0220 08:09:37.679994 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727\": container with ID starting with 37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727 not found: ID does not exist" containerID="37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.680037 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727"} err="failed to get container status \"37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727\": rpc error: code = NotFound desc = could not find container \"37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727\": container with ID starting with 37d889ae6047df55260b1c8d1e60efd8e2c713cc29261563a7b5f099de30c727 not found: ID does not exist" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.680067 4492 scope.go:117] "RemoveContainer" containerID="52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178" Feb 20 08:09:37 crc kubenswrapper[4492]: E0220 08:09:37.680415 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178\": container with ID starting with 52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178 not found: ID does not exist" containerID="52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.680439 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178"} err="failed to get container status \"52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178\": rpc error: code = NotFound desc = could not find container \"52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178\": container with ID starting with 52997309d3011170bc5bc80a8cfb8418286b38c796a9f48b3637f6fbf7645178 not found: ID does not exist" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.680459 4492 scope.go:117] "RemoveContainer" containerID="f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c" Feb 20 08:09:37 crc kubenswrapper[4492]: E0220 08:09:37.680728 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c\": container with ID starting with f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c not found: ID does not exist" containerID="f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c" Feb 20 08:09:37 crc kubenswrapper[4492]: I0220 08:09:37.680751 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c"} err="failed to get container status \"f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c\": rpc error: code = NotFound desc = could not find container \"f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c\": container with ID starting with f561fada322692a166cfc86522dd9fc0d2be7ec186307222c7017c67e14f106c not found: ID does not exist" Feb 20 08:09:39 crc kubenswrapper[4492]: I0220 08:09:39.565955 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" path="/var/lib/kubelet/pods/469df4a7-7e1b-434c-9605-f79bb68c196c/volumes" Feb 20 08:10:39 crc kubenswrapper[4492]: I0220 08:10:39.311738 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:10:39 crc kubenswrapper[4492]: I0220 08:10:39.312240 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:11:09 crc kubenswrapper[4492]: I0220 08:11:09.311158 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:11:09 crc kubenswrapper[4492]: I0220 08:11:09.312344 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.254351 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhmv"] Feb 20 08:11:15 crc kubenswrapper[4492]: E0220 08:11:15.257034 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="registry-server" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.257139 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="registry-server" Feb 20 08:11:15 crc kubenswrapper[4492]: E0220 08:11:15.257235 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="extract-utilities" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.257284 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="extract-utilities" Feb 20 08:11:15 crc kubenswrapper[4492]: E0220 08:11:15.257349 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="extract-content" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.257398 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="extract-content" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.257690 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="469df4a7-7e1b-434c-9605-f79bb68c196c" containerName="registry-server" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.263851 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.321455 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhmv"] Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.391853 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-utilities\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.392297 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-catalog-content\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.392332 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4dst\" (UniqueName: \"kubernetes.io/projected/e24e3dca-96f5-4840-acaa-88497045422f-kube-api-access-x4dst\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.494114 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-catalog-content\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.494168 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4dst\" (UniqueName: \"kubernetes.io/projected/e24e3dca-96f5-4840-acaa-88497045422f-kube-api-access-x4dst\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.494359 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-utilities\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.494812 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-catalog-content\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.494917 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-utilities\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.518774 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4dst\" (UniqueName: \"kubernetes.io/projected/e24e3dca-96f5-4840-acaa-88497045422f-kube-api-access-x4dst\") pod \"redhat-marketplace-qzhmv\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:15 crc kubenswrapper[4492]: I0220 08:11:15.583824 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:16 crc kubenswrapper[4492]: I0220 08:11:16.063953 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhmv"] Feb 20 08:11:16 crc kubenswrapper[4492]: I0220 08:11:16.508696 4492 generic.go:334] "Generic (PLEG): container finished" podID="e24e3dca-96f5-4840-acaa-88497045422f" containerID="b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398" exitCode=0 Feb 20 08:11:16 crc kubenswrapper[4492]: I0220 08:11:16.508802 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhmv" event={"ID":"e24e3dca-96f5-4840-acaa-88497045422f","Type":"ContainerDied","Data":"b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398"} Feb 20 08:11:16 crc kubenswrapper[4492]: I0220 08:11:16.508932 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhmv" event={"ID":"e24e3dca-96f5-4840-acaa-88497045422f","Type":"ContainerStarted","Data":"88e5d64912b125b8e8b6cabbbef8a7bf009e11d56a26868e27f5913a4d853444"} Feb 20 08:11:17 crc kubenswrapper[4492]: I0220 08:11:17.520667 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhmv" event={"ID":"e24e3dca-96f5-4840-acaa-88497045422f","Type":"ContainerStarted","Data":"3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7"} Feb 20 08:11:18 crc kubenswrapper[4492]: I0220 08:11:18.529364 4492 generic.go:334] "Generic (PLEG): container finished" podID="e24e3dca-96f5-4840-acaa-88497045422f" containerID="3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7" exitCode=0 Feb 20 08:11:18 crc kubenswrapper[4492]: I0220 08:11:18.529467 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhmv" event={"ID":"e24e3dca-96f5-4840-acaa-88497045422f","Type":"ContainerDied","Data":"3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7"} Feb 20 08:11:19 crc kubenswrapper[4492]: I0220 08:11:19.539231 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhmv" event={"ID":"e24e3dca-96f5-4840-acaa-88497045422f","Type":"ContainerStarted","Data":"70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294"} Feb 20 08:11:19 crc kubenswrapper[4492]: I0220 08:11:19.565179 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qzhmv" podStartSLOduration=2.034978362 podStartE2EDuration="4.565161592s" podCreationTimestamp="2026-02-20 08:11:15 +0000 UTC" firstStartedPulling="2026-02-20 08:11:16.511170875 +0000 UTC m=+5433.282459853" lastFinishedPulling="2026-02-20 08:11:19.041354105 +0000 UTC m=+5435.812643083" observedRunningTime="2026-02-20 08:11:19.556793902 +0000 UTC m=+5436.328082880" watchObservedRunningTime="2026-02-20 08:11:19.565161592 +0000 UTC m=+5436.336450569" Feb 20 08:11:25 crc kubenswrapper[4492]: I0220 08:11:25.585033 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:25 crc kubenswrapper[4492]: I0220 08:11:25.585734 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:25 crc kubenswrapper[4492]: I0220 08:11:25.625870 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:25 crc kubenswrapper[4492]: I0220 08:11:25.666905 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:25 crc kubenswrapper[4492]: I0220 08:11:25.863278 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhmv"] Feb 20 08:11:27 crc kubenswrapper[4492]: I0220 08:11:27.619529 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qzhmv" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="registry-server" containerID="cri-o://70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294" gracePeriod=2 Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.075857 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.244215 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-catalog-content\") pod \"e24e3dca-96f5-4840-acaa-88497045422f\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.244301 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4dst\" (UniqueName: \"kubernetes.io/projected/e24e3dca-96f5-4840-acaa-88497045422f-kube-api-access-x4dst\") pod \"e24e3dca-96f5-4840-acaa-88497045422f\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.244444 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-utilities\") pod \"e24e3dca-96f5-4840-acaa-88497045422f\" (UID: \"e24e3dca-96f5-4840-acaa-88497045422f\") " Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.245156 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-utilities" (OuterVolumeSpecName: "utilities") pod "e24e3dca-96f5-4840-acaa-88497045422f" (UID: "e24e3dca-96f5-4840-acaa-88497045422f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.250099 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e24e3dca-96f5-4840-acaa-88497045422f-kube-api-access-x4dst" (OuterVolumeSpecName: "kube-api-access-x4dst") pod "e24e3dca-96f5-4840-acaa-88497045422f" (UID: "e24e3dca-96f5-4840-acaa-88497045422f"). InnerVolumeSpecName "kube-api-access-x4dst". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.262572 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e24e3dca-96f5-4840-acaa-88497045422f" (UID: "e24e3dca-96f5-4840-acaa-88497045422f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.346214 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.346240 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e24e3dca-96f5-4840-acaa-88497045422f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.346252 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4dst\" (UniqueName: \"kubernetes.io/projected/e24e3dca-96f5-4840-acaa-88497045422f-kube-api-access-x4dst\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.630935 4492 generic.go:334] "Generic (PLEG): container finished" podID="e24e3dca-96f5-4840-acaa-88497045422f" containerID="70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294" exitCode=0 Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.631016 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzhmv" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.631007 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhmv" event={"ID":"e24e3dca-96f5-4840-acaa-88497045422f","Type":"ContainerDied","Data":"70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294"} Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.631956 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzhmv" event={"ID":"e24e3dca-96f5-4840-acaa-88497045422f","Type":"ContainerDied","Data":"88e5d64912b125b8e8b6cabbbef8a7bf009e11d56a26868e27f5913a4d853444"} Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.631982 4492 scope.go:117] "RemoveContainer" containerID="70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.657851 4492 scope.go:117] "RemoveContainer" containerID="3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.670300 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhmv"] Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.682016 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzhmv"] Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.689465 4492 scope.go:117] "RemoveContainer" containerID="b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.718560 4492 scope.go:117] "RemoveContainer" containerID="70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294" Feb 20 08:11:28 crc kubenswrapper[4492]: E0220 08:11:28.718947 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294\": container with ID starting with 70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294 not found: ID does not exist" containerID="70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.718980 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294"} err="failed to get container status \"70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294\": rpc error: code = NotFound desc = could not find container \"70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294\": container with ID starting with 70ef8f0949e43e1c6db86363f13fe68130a63bee0ec6a061ef58c86c8fdfe294 not found: ID does not exist" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.719008 4492 scope.go:117] "RemoveContainer" containerID="3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7" Feb 20 08:11:28 crc kubenswrapper[4492]: E0220 08:11:28.719375 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7\": container with ID starting with 3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7 not found: ID does not exist" containerID="3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.719400 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7"} err="failed to get container status \"3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7\": rpc error: code = NotFound desc = could not find container \"3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7\": container with ID starting with 3032651ba5e82227490515b49c07d936b2fcc4a2e138fa1c5aa3f9f84fe0adf7 not found: ID does not exist" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.719417 4492 scope.go:117] "RemoveContainer" containerID="b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398" Feb 20 08:11:28 crc kubenswrapper[4492]: E0220 08:11:28.719817 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398\": container with ID starting with b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398 not found: ID does not exist" containerID="b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398" Feb 20 08:11:28 crc kubenswrapper[4492]: I0220 08:11:28.719840 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398"} err="failed to get container status \"b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398\": rpc error: code = NotFound desc = could not find container \"b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398\": container with ID starting with b5abd13ffc056a073d37e15ff66e151c0f781cdd3ffe93da561d12af4bbaf398 not found: ID does not exist" Feb 20 08:11:29 crc kubenswrapper[4492]: I0220 08:11:29.569744 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e24e3dca-96f5-4840-acaa-88497045422f" path="/var/lib/kubelet/pods/e24e3dca-96f5-4840-acaa-88497045422f/volumes" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.270621 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nqg8b"] Feb 20 08:11:31 crc kubenswrapper[4492]: E0220 08:11:31.271220 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="extract-content" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.271236 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="extract-content" Feb 20 08:11:31 crc kubenswrapper[4492]: E0220 08:11:31.271256 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="registry-server" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.271264 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="registry-server" Feb 20 08:11:31 crc kubenswrapper[4492]: E0220 08:11:31.271281 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="extract-utilities" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.271287 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="extract-utilities" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.271524 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e24e3dca-96f5-4840-acaa-88497045422f" containerName="registry-server" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.273046 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.281392 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nqg8b"] Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.405852 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-catalog-content\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.405927 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-utilities\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.406021 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkzmc\" (UniqueName: \"kubernetes.io/projected/83447257-16fe-4130-950f-2434a7fa2e48-kube-api-access-kkzmc\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.507983 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkzmc\" (UniqueName: \"kubernetes.io/projected/83447257-16fe-4130-950f-2434a7fa2e48-kube-api-access-kkzmc\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.508710 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-catalog-content\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.508768 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-utilities\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.509209 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-utilities\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.509282 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-catalog-content\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.525243 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkzmc\" (UniqueName: \"kubernetes.io/projected/83447257-16fe-4130-950f-2434a7fa2e48-kube-api-access-kkzmc\") pod \"redhat-operators-nqg8b\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.590277 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:31 crc kubenswrapper[4492]: I0220 08:11:31.987797 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nqg8b"] Feb 20 08:11:32 crc kubenswrapper[4492]: I0220 08:11:32.689410 4492 generic.go:334] "Generic (PLEG): container finished" podID="83447257-16fe-4130-950f-2434a7fa2e48" containerID="3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259" exitCode=0 Feb 20 08:11:32 crc kubenswrapper[4492]: I0220 08:11:32.689533 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqg8b" event={"ID":"83447257-16fe-4130-950f-2434a7fa2e48","Type":"ContainerDied","Data":"3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259"} Feb 20 08:11:32 crc kubenswrapper[4492]: I0220 08:11:32.689746 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqg8b" event={"ID":"83447257-16fe-4130-950f-2434a7fa2e48","Type":"ContainerStarted","Data":"c2faf3a871faa0f57ac63a1e1c123a92bc7e2a96d202226529be3d7baaea5e1c"} Feb 20 08:11:33 crc kubenswrapper[4492]: I0220 08:11:33.700918 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqg8b" event={"ID":"83447257-16fe-4130-950f-2434a7fa2e48","Type":"ContainerStarted","Data":"d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a"} Feb 20 08:11:36 crc kubenswrapper[4492]: I0220 08:11:36.741352 4492 generic.go:334] "Generic (PLEG): container finished" podID="83447257-16fe-4130-950f-2434a7fa2e48" containerID="d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a" exitCode=0 Feb 20 08:11:36 crc kubenswrapper[4492]: I0220 08:11:36.741439 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqg8b" event={"ID":"83447257-16fe-4130-950f-2434a7fa2e48","Type":"ContainerDied","Data":"d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a"} Feb 20 08:11:37 crc kubenswrapper[4492]: I0220 08:11:37.753364 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqg8b" event={"ID":"83447257-16fe-4130-950f-2434a7fa2e48","Type":"ContainerStarted","Data":"cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51"} Feb 20 08:11:37 crc kubenswrapper[4492]: I0220 08:11:37.782421 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nqg8b" podStartSLOduration=2.276472557 podStartE2EDuration="6.782404194s" podCreationTimestamp="2026-02-20 08:11:31 +0000 UTC" firstStartedPulling="2026-02-20 08:11:32.691397972 +0000 UTC m=+5449.462686950" lastFinishedPulling="2026-02-20 08:11:37.197329609 +0000 UTC m=+5453.968618587" observedRunningTime="2026-02-20 08:11:37.776680078 +0000 UTC m=+5454.547969056" watchObservedRunningTime="2026-02-20 08:11:37.782404194 +0000 UTC m=+5454.553693171" Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.311122 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.311535 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.311591 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.312222 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"53c62c1dc0f3353ec3654a5971e2430befc77fd708a235e5633ff86ec1759add"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.312277 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://53c62c1dc0f3353ec3654a5971e2430befc77fd708a235e5633ff86ec1759add" gracePeriod=600 Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.772827 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="53c62c1dc0f3353ec3654a5971e2430befc77fd708a235e5633ff86ec1759add" exitCode=0 Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.772927 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"53c62c1dc0f3353ec3654a5971e2430befc77fd708a235e5633ff86ec1759add"} Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.773193 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449"} Feb 20 08:11:39 crc kubenswrapper[4492]: I0220 08:11:39.773223 4492 scope.go:117] "RemoveContainer" containerID="60c1789d9be10f0f4e93b6bdeb5a8dbe42f2b484817a8797e4676030f2db5b26" Feb 20 08:11:41 crc kubenswrapper[4492]: I0220 08:11:41.590899 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:41 crc kubenswrapper[4492]: I0220 08:11:41.591311 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:42 crc kubenswrapper[4492]: I0220 08:11:42.630800 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nqg8b" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="registry-server" probeResult="failure" output=< Feb 20 08:11:42 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:11:42 crc kubenswrapper[4492]: > Feb 20 08:11:51 crc kubenswrapper[4492]: I0220 08:11:51.660644 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:51 crc kubenswrapper[4492]: I0220 08:11:51.722940 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:51 crc kubenswrapper[4492]: I0220 08:11:51.927368 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nqg8b"] Feb 20 08:11:52 crc kubenswrapper[4492]: I0220 08:11:52.907909 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nqg8b" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="registry-server" containerID="cri-o://cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51" gracePeriod=2 Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.487687 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.585176 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-catalog-content\") pod \"83447257-16fe-4130-950f-2434a7fa2e48\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.585496 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkzmc\" (UniqueName: \"kubernetes.io/projected/83447257-16fe-4130-950f-2434a7fa2e48-kube-api-access-kkzmc\") pod \"83447257-16fe-4130-950f-2434a7fa2e48\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.585738 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-utilities\") pod \"83447257-16fe-4130-950f-2434a7fa2e48\" (UID: \"83447257-16fe-4130-950f-2434a7fa2e48\") " Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.587179 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-utilities" (OuterVolumeSpecName: "utilities") pod "83447257-16fe-4130-950f-2434a7fa2e48" (UID: "83447257-16fe-4130-950f-2434a7fa2e48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.600294 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83447257-16fe-4130-950f-2434a7fa2e48-kube-api-access-kkzmc" (OuterVolumeSpecName: "kube-api-access-kkzmc") pod "83447257-16fe-4130-950f-2434a7fa2e48" (UID: "83447257-16fe-4130-950f-2434a7fa2e48"). InnerVolumeSpecName "kube-api-access-kkzmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.684665 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83447257-16fe-4130-950f-2434a7fa2e48" (UID: "83447257-16fe-4130-950f-2434a7fa2e48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.689339 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.689382 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83447257-16fe-4130-950f-2434a7fa2e48-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.689395 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkzmc\" (UniqueName: \"kubernetes.io/projected/83447257-16fe-4130-950f-2434a7fa2e48-kube-api-access-kkzmc\") on node \"crc\" DevicePath \"\"" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.917889 4492 generic.go:334] "Generic (PLEG): container finished" podID="83447257-16fe-4130-950f-2434a7fa2e48" containerID="cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51" exitCode=0 Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.917957 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nqg8b" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.917959 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqg8b" event={"ID":"83447257-16fe-4130-950f-2434a7fa2e48","Type":"ContainerDied","Data":"cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51"} Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.918322 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nqg8b" event={"ID":"83447257-16fe-4130-950f-2434a7fa2e48","Type":"ContainerDied","Data":"c2faf3a871faa0f57ac63a1e1c123a92bc7e2a96d202226529be3d7baaea5e1c"} Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.918343 4492 scope.go:117] "RemoveContainer" containerID="cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.946400 4492 scope.go:117] "RemoveContainer" containerID="d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a" Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.955611 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nqg8b"] Feb 20 08:11:53 crc kubenswrapper[4492]: I0220 08:11:53.964039 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nqg8b"] Feb 20 08:11:54 crc kubenswrapper[4492]: I0220 08:11:53.998016 4492 scope.go:117] "RemoveContainer" containerID="3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259" Feb 20 08:11:54 crc kubenswrapper[4492]: I0220 08:11:54.031270 4492 scope.go:117] "RemoveContainer" containerID="cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51" Feb 20 08:11:54 crc kubenswrapper[4492]: E0220 08:11:54.031706 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51\": container with ID starting with cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51 not found: ID does not exist" containerID="cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51" Feb 20 08:11:54 crc kubenswrapper[4492]: I0220 08:11:54.031733 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51"} err="failed to get container status \"cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51\": rpc error: code = NotFound desc = could not find container \"cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51\": container with ID starting with cf3ec150b2073e42d1a5131d59fe757a6a6da9e0cca09f2bf817f6bf568e1e51 not found: ID does not exist" Feb 20 08:11:54 crc kubenswrapper[4492]: I0220 08:11:54.031754 4492 scope.go:117] "RemoveContainer" containerID="d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a" Feb 20 08:11:54 crc kubenswrapper[4492]: E0220 08:11:54.032103 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a\": container with ID starting with d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a not found: ID does not exist" containerID="d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a" Feb 20 08:11:54 crc kubenswrapper[4492]: I0220 08:11:54.032127 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a"} err="failed to get container status \"d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a\": rpc error: code = NotFound desc = could not find container \"d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a\": container with ID starting with d4417442642456e86332aeb964354fee681ef0590e7787281ada6c2238cd210a not found: ID does not exist" Feb 20 08:11:54 crc kubenswrapper[4492]: I0220 08:11:54.032141 4492 scope.go:117] "RemoveContainer" containerID="3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259" Feb 20 08:11:54 crc kubenswrapper[4492]: E0220 08:11:54.032412 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259\": container with ID starting with 3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259 not found: ID does not exist" containerID="3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259" Feb 20 08:11:54 crc kubenswrapper[4492]: I0220 08:11:54.032435 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259"} err="failed to get container status \"3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259\": rpc error: code = NotFound desc = could not find container \"3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259\": container with ID starting with 3820ab28b4534e95c29ec687446bbee2180ab7bded7da9e4fc731e868f3ad259 not found: ID does not exist" Feb 20 08:11:55 crc kubenswrapper[4492]: I0220 08:11:55.566213 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83447257-16fe-4130-950f-2434a7fa2e48" path="/var/lib/kubelet/pods/83447257-16fe-4130-950f-2434a7fa2e48/volumes" Feb 20 08:13:39 crc kubenswrapper[4492]: I0220 08:13:39.311779 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:13:39 crc kubenswrapper[4492]: I0220 08:13:39.314002 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:14:09 crc kubenswrapper[4492]: I0220 08:14:09.311754 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:14:09 crc kubenswrapper[4492]: I0220 08:14:09.312228 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.752582 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-drjc5"] Feb 20 08:14:20 crc kubenswrapper[4492]: E0220 08:14:20.753500 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="extract-utilities" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.753519 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="extract-utilities" Feb 20 08:14:20 crc kubenswrapper[4492]: E0220 08:14:20.753542 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="extract-content" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.753548 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="extract-content" Feb 20 08:14:20 crc kubenswrapper[4492]: E0220 08:14:20.753562 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="registry-server" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.753568 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="registry-server" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.753768 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="83447257-16fe-4130-950f-2434a7fa2e48" containerName="registry-server" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.755247 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.776350 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-drjc5"] Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.931161 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c5ff\" (UniqueName: \"kubernetes.io/projected/e2775036-25a6-4a76-86c4-5435e7c89dab-kube-api-access-5c5ff\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.931206 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-utilities\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:20 crc kubenswrapper[4492]: I0220 08:14:20.931377 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-catalog-content\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.032894 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c5ff\" (UniqueName: \"kubernetes.io/projected/e2775036-25a6-4a76-86c4-5435e7c89dab-kube-api-access-5c5ff\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.033195 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-utilities\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.033685 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-utilities\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.033838 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-catalog-content\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.034095 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-catalog-content\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.066268 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c5ff\" (UniqueName: \"kubernetes.io/projected/e2775036-25a6-4a76-86c4-5435e7c89dab-kube-api-access-5c5ff\") pod \"community-operators-drjc5\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.075528 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:21 crc kubenswrapper[4492]: I0220 08:14:21.657345 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-drjc5"] Feb 20 08:14:22 crc kubenswrapper[4492]: I0220 08:14:22.239108 4492 generic.go:334] "Generic (PLEG): container finished" podID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerID="bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9" exitCode=0 Feb 20 08:14:22 crc kubenswrapper[4492]: I0220 08:14:22.239173 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drjc5" event={"ID":"e2775036-25a6-4a76-86c4-5435e7c89dab","Type":"ContainerDied","Data":"bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9"} Feb 20 08:14:22 crc kubenswrapper[4492]: I0220 08:14:22.239357 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drjc5" event={"ID":"e2775036-25a6-4a76-86c4-5435e7c89dab","Type":"ContainerStarted","Data":"ad36ab3d901fe67787a5200078c635c39a855b5f3eb86912776fff3d82711dbf"} Feb 20 08:14:23 crc kubenswrapper[4492]: I0220 08:14:23.249315 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drjc5" event={"ID":"e2775036-25a6-4a76-86c4-5435e7c89dab","Type":"ContainerStarted","Data":"1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00"} Feb 20 08:14:24 crc kubenswrapper[4492]: I0220 08:14:24.259401 4492 generic.go:334] "Generic (PLEG): container finished" podID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerID="1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00" exitCode=0 Feb 20 08:14:24 crc kubenswrapper[4492]: I0220 08:14:24.259447 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drjc5" event={"ID":"e2775036-25a6-4a76-86c4-5435e7c89dab","Type":"ContainerDied","Data":"1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00"} Feb 20 08:14:25 crc kubenswrapper[4492]: I0220 08:14:25.269586 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drjc5" event={"ID":"e2775036-25a6-4a76-86c4-5435e7c89dab","Type":"ContainerStarted","Data":"58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa"} Feb 20 08:14:25 crc kubenswrapper[4492]: I0220 08:14:25.288286 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-drjc5" podStartSLOduration=2.782432933 podStartE2EDuration="5.288266605s" podCreationTimestamp="2026-02-20 08:14:20 +0000 UTC" firstStartedPulling="2026-02-20 08:14:22.240925186 +0000 UTC m=+5619.012214165" lastFinishedPulling="2026-02-20 08:14:24.746758858 +0000 UTC m=+5621.518047837" observedRunningTime="2026-02-20 08:14:25.285026864 +0000 UTC m=+5622.056315843" watchObservedRunningTime="2026-02-20 08:14:25.288266605 +0000 UTC m=+5622.059555582" Feb 20 08:14:31 crc kubenswrapper[4492]: I0220 08:14:31.076664 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:31 crc kubenswrapper[4492]: I0220 08:14:31.077364 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:31 crc kubenswrapper[4492]: I0220 08:14:31.117236 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:31 crc kubenswrapper[4492]: I0220 08:14:31.354410 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:32 crc kubenswrapper[4492]: I0220 08:14:32.142858 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-drjc5"] Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.335915 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-drjc5" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="registry-server" containerID="cri-o://58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa" gracePeriod=2 Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.771115 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.797152 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-catalog-content\") pod \"e2775036-25a6-4a76-86c4-5435e7c89dab\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.797442 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-utilities\") pod \"e2775036-25a6-4a76-86c4-5435e7c89dab\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.797570 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c5ff\" (UniqueName: \"kubernetes.io/projected/e2775036-25a6-4a76-86c4-5435e7c89dab-kube-api-access-5c5ff\") pod \"e2775036-25a6-4a76-86c4-5435e7c89dab\" (UID: \"e2775036-25a6-4a76-86c4-5435e7c89dab\") " Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.798087 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-utilities" (OuterVolumeSpecName: "utilities") pod "e2775036-25a6-4a76-86c4-5435e7c89dab" (UID: "e2775036-25a6-4a76-86c4-5435e7c89dab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.809415 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2775036-25a6-4a76-86c4-5435e7c89dab-kube-api-access-5c5ff" (OuterVolumeSpecName: "kube-api-access-5c5ff") pod "e2775036-25a6-4a76-86c4-5435e7c89dab" (UID: "e2775036-25a6-4a76-86c4-5435e7c89dab"). InnerVolumeSpecName "kube-api-access-5c5ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.844983 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2775036-25a6-4a76-86c4-5435e7c89dab" (UID: "e2775036-25a6-4a76-86c4-5435e7c89dab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.902312 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.902355 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2775036-25a6-4a76-86c4-5435e7c89dab-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:14:33 crc kubenswrapper[4492]: I0220 08:14:33.902371 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c5ff\" (UniqueName: \"kubernetes.io/projected/e2775036-25a6-4a76-86c4-5435e7c89dab-kube-api-access-5c5ff\") on node \"crc\" DevicePath \"\"" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.346888 4492 generic.go:334] "Generic (PLEG): container finished" podID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerID="58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa" exitCode=0 Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.346954 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drjc5" event={"ID":"e2775036-25a6-4a76-86c4-5435e7c89dab","Type":"ContainerDied","Data":"58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa"} Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.346964 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-drjc5" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.347005 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-drjc5" event={"ID":"e2775036-25a6-4a76-86c4-5435e7c89dab","Type":"ContainerDied","Data":"ad36ab3d901fe67787a5200078c635c39a855b5f3eb86912776fff3d82711dbf"} Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.347029 4492 scope.go:117] "RemoveContainer" containerID="58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.375272 4492 scope.go:117] "RemoveContainer" containerID="1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.388094 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-drjc5"] Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.394939 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-drjc5"] Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.398788 4492 scope.go:117] "RemoveContainer" containerID="bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.431973 4492 scope.go:117] "RemoveContainer" containerID="58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa" Feb 20 08:14:34 crc kubenswrapper[4492]: E0220 08:14:34.432548 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa\": container with ID starting with 58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa not found: ID does not exist" containerID="58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.432580 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa"} err="failed to get container status \"58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa\": rpc error: code = NotFound desc = could not find container \"58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa\": container with ID starting with 58712c873907a7b9dc111b25d268a355c90aaabeb7faeac6e80ee4673f09e4aa not found: ID does not exist" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.432598 4492 scope.go:117] "RemoveContainer" containerID="1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00" Feb 20 08:14:34 crc kubenswrapper[4492]: E0220 08:14:34.432934 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00\": container with ID starting with 1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00 not found: ID does not exist" containerID="1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.432979 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00"} err="failed to get container status \"1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00\": rpc error: code = NotFound desc = could not find container \"1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00\": container with ID starting with 1a472cdd29414a42f315b7fd6bb1182744eb974680ef4abbb71939a40f484f00 not found: ID does not exist" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.433008 4492 scope.go:117] "RemoveContainer" containerID="bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9" Feb 20 08:14:34 crc kubenswrapper[4492]: E0220 08:14:34.433331 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9\": container with ID starting with bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9 not found: ID does not exist" containerID="bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9" Feb 20 08:14:34 crc kubenswrapper[4492]: I0220 08:14:34.433350 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9"} err="failed to get container status \"bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9\": rpc error: code = NotFound desc = could not find container \"bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9\": container with ID starting with bb082d48437aa853f915e3ca15fe037a5326fa3b17beb8b6c9c43bf4f8d90ba9 not found: ID does not exist" Feb 20 08:14:35 crc kubenswrapper[4492]: I0220 08:14:35.567555 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" path="/var/lib/kubelet/pods/e2775036-25a6-4a76-86c4-5435e7c89dab/volumes" Feb 20 08:14:39 crc kubenswrapper[4492]: I0220 08:14:39.311313 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:14:39 crc kubenswrapper[4492]: I0220 08:14:39.312005 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:14:39 crc kubenswrapper[4492]: I0220 08:14:39.312057 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:14:39 crc kubenswrapper[4492]: I0220 08:14:39.312591 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:14:39 crc kubenswrapper[4492]: I0220 08:14:39.312648 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" gracePeriod=600 Feb 20 08:14:39 crc kubenswrapper[4492]: E0220 08:14:39.433069 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:14:40 crc kubenswrapper[4492]: I0220 08:14:40.400213 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" exitCode=0 Feb 20 08:14:40 crc kubenswrapper[4492]: I0220 08:14:40.400318 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449"} Feb 20 08:14:40 crc kubenswrapper[4492]: I0220 08:14:40.400647 4492 scope.go:117] "RemoveContainer" containerID="53c62c1dc0f3353ec3654a5971e2430befc77fd708a235e5633ff86ec1759add" Feb 20 08:14:40 crc kubenswrapper[4492]: I0220 08:14:40.401347 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:14:40 crc kubenswrapper[4492]: E0220 08:14:40.401757 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:14:51 crc kubenswrapper[4492]: I0220 08:14:51.556938 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:14:51 crc kubenswrapper[4492]: E0220 08:14:51.557900 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.171001 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv"] Feb 20 08:15:00 crc kubenswrapper[4492]: E0220 08:15:00.172364 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="registry-server" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.172386 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="registry-server" Feb 20 08:15:00 crc kubenswrapper[4492]: E0220 08:15:00.172498 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="extract-content" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.172507 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="extract-content" Feb 20 08:15:00 crc kubenswrapper[4492]: E0220 08:15:00.172526 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="extract-utilities" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.172534 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="extract-utilities" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.172831 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2775036-25a6-4a76-86c4-5435e7c89dab" containerName="registry-server" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.173843 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.206269 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.206691 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.271320 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv"] Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.323731 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wfzk\" (UniqueName: \"kubernetes.io/projected/64892d2a-51dd-4ae6-8936-3c9591b32884-kube-api-access-5wfzk\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.323790 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64892d2a-51dd-4ae6-8936-3c9591b32884-secret-volume\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.323865 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64892d2a-51dd-4ae6-8936-3c9591b32884-config-volume\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.426213 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wfzk\" (UniqueName: \"kubernetes.io/projected/64892d2a-51dd-4ae6-8936-3c9591b32884-kube-api-access-5wfzk\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.426274 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64892d2a-51dd-4ae6-8936-3c9591b32884-secret-volume\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.426404 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64892d2a-51dd-4ae6-8936-3c9591b32884-config-volume\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.427347 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64892d2a-51dd-4ae6-8936-3c9591b32884-config-volume\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.438827 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64892d2a-51dd-4ae6-8936-3c9591b32884-secret-volume\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.446326 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wfzk\" (UniqueName: \"kubernetes.io/projected/64892d2a-51dd-4ae6-8936-3c9591b32884-kube-api-access-5wfzk\") pod \"collect-profiles-29526255-8tsjv\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.495788 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:00 crc kubenswrapper[4492]: I0220 08:15:00.927001 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv"] Feb 20 08:15:01 crc kubenswrapper[4492]: I0220 08:15:01.610699 4492 generic.go:334] "Generic (PLEG): container finished" podID="64892d2a-51dd-4ae6-8936-3c9591b32884" containerID="b18eee7fda7f7686f5eecca57bf90554594eb9c243edfdf9ac19c79598cb5a8e" exitCode=0 Feb 20 08:15:01 crc kubenswrapper[4492]: I0220 08:15:01.610891 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" event={"ID":"64892d2a-51dd-4ae6-8936-3c9591b32884","Type":"ContainerDied","Data":"b18eee7fda7f7686f5eecca57bf90554594eb9c243edfdf9ac19c79598cb5a8e"} Feb 20 08:15:01 crc kubenswrapper[4492]: I0220 08:15:01.611978 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" event={"ID":"64892d2a-51dd-4ae6-8936-3c9591b32884","Type":"ContainerStarted","Data":"ff43e9d838d36c3aecdbee45929e74232842cb43723e1e8272249e84a43304e8"} Feb 20 08:15:02 crc kubenswrapper[4492]: I0220 08:15:02.936756 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:02 crc kubenswrapper[4492]: I0220 08:15:02.989150 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wfzk\" (UniqueName: \"kubernetes.io/projected/64892d2a-51dd-4ae6-8936-3c9591b32884-kube-api-access-5wfzk\") pod \"64892d2a-51dd-4ae6-8936-3c9591b32884\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " Feb 20 08:15:02 crc kubenswrapper[4492]: I0220 08:15:02.989262 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64892d2a-51dd-4ae6-8936-3c9591b32884-secret-volume\") pod \"64892d2a-51dd-4ae6-8936-3c9591b32884\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " Feb 20 08:15:02 crc kubenswrapper[4492]: I0220 08:15:02.989419 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64892d2a-51dd-4ae6-8936-3c9591b32884-config-volume\") pod \"64892d2a-51dd-4ae6-8936-3c9591b32884\" (UID: \"64892d2a-51dd-4ae6-8936-3c9591b32884\") " Feb 20 08:15:02 crc kubenswrapper[4492]: I0220 08:15:02.990502 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64892d2a-51dd-4ae6-8936-3c9591b32884-config-volume" (OuterVolumeSpecName: "config-volume") pod "64892d2a-51dd-4ae6-8936-3c9591b32884" (UID: "64892d2a-51dd-4ae6-8936-3c9591b32884"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:15:02 crc kubenswrapper[4492]: I0220 08:15:02.997611 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64892d2a-51dd-4ae6-8936-3c9591b32884-kube-api-access-5wfzk" (OuterVolumeSpecName: "kube-api-access-5wfzk") pod "64892d2a-51dd-4ae6-8936-3c9591b32884" (UID: "64892d2a-51dd-4ae6-8936-3c9591b32884"). InnerVolumeSpecName "kube-api-access-5wfzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.006723 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64892d2a-51dd-4ae6-8936-3c9591b32884-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "64892d2a-51dd-4ae6-8936-3c9591b32884" (UID: "64892d2a-51dd-4ae6-8936-3c9591b32884"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.090960 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wfzk\" (UniqueName: \"kubernetes.io/projected/64892d2a-51dd-4ae6-8936-3c9591b32884-kube-api-access-5wfzk\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.090990 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64892d2a-51dd-4ae6-8936-3c9591b32884-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.091000 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64892d2a-51dd-4ae6-8936-3c9591b32884-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.562314 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:15:03 crc kubenswrapper[4492]: E0220 08:15:03.563019 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.633761 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" event={"ID":"64892d2a-51dd-4ae6-8936-3c9591b32884","Type":"ContainerDied","Data":"ff43e9d838d36c3aecdbee45929e74232842cb43723e1e8272249e84a43304e8"} Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.633949 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv" Feb 20 08:15:03 crc kubenswrapper[4492]: I0220 08:15:03.633811 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff43e9d838d36c3aecdbee45929e74232842cb43723e1e8272249e84a43304e8" Feb 20 08:15:04 crc kubenswrapper[4492]: I0220 08:15:04.053359 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm"] Feb 20 08:15:04 crc kubenswrapper[4492]: I0220 08:15:04.060659 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526210-mk4qm"] Feb 20 08:15:05 crc kubenswrapper[4492]: I0220 08:15:05.566770 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01f19eb-c60f-4612-8a50-a78564ab6654" path="/var/lib/kubelet/pods/a01f19eb-c60f-4612-8a50-a78564ab6654/volumes" Feb 20 08:15:09 crc kubenswrapper[4492]: I0220 08:15:09.311123 4492 scope.go:117] "RemoveContainer" containerID="8eee68dfa83451ae4b028554f3a937821af25e2d4c060d765336df54f980093e" Feb 20 08:15:14 crc kubenswrapper[4492]: I0220 08:15:14.557444 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:15:14 crc kubenswrapper[4492]: E0220 08:15:14.559023 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:15:26 crc kubenswrapper[4492]: I0220 08:15:26.558107 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:15:26 crc kubenswrapper[4492]: E0220 08:15:26.559055 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:15:41 crc kubenswrapper[4492]: I0220 08:15:41.556812 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:15:41 crc kubenswrapper[4492]: E0220 08:15:41.557560 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:15:52 crc kubenswrapper[4492]: I0220 08:15:52.558752 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:15:52 crc kubenswrapper[4492]: E0220 08:15:52.560487 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:16:07 crc kubenswrapper[4492]: I0220 08:16:07.557087 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:16:07 crc kubenswrapper[4492]: E0220 08:16:07.557783 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:16:20 crc kubenswrapper[4492]: I0220 08:16:20.557285 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:16:20 crc kubenswrapper[4492]: E0220 08:16:20.558178 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:16:34 crc kubenswrapper[4492]: I0220 08:16:34.557282 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:16:34 crc kubenswrapper[4492]: E0220 08:16:34.558030 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:16:37 crc kubenswrapper[4492]: E0220 08:16:37.573258 4492 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.25:52400->192.168.26.25:44007: write tcp 192.168.26.25:52400->192.168.26.25:44007: write: broken pipe Feb 20 08:16:48 crc kubenswrapper[4492]: I0220 08:16:48.557714 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:16:48 crc kubenswrapper[4492]: E0220 08:16:48.558734 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:17:00 crc kubenswrapper[4492]: I0220 08:17:00.557334 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:17:00 crc kubenswrapper[4492]: E0220 08:17:00.558222 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:17:12 crc kubenswrapper[4492]: I0220 08:17:12.559067 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:17:12 crc kubenswrapper[4492]: E0220 08:17:12.560457 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:17:27 crc kubenswrapper[4492]: I0220 08:17:27.557288 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:17:27 crc kubenswrapper[4492]: E0220 08:17:27.558250 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:17:42 crc kubenswrapper[4492]: I0220 08:17:42.557090 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:17:42 crc kubenswrapper[4492]: E0220 08:17:42.558232 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:17:55 crc kubenswrapper[4492]: I0220 08:17:55.558566 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:17:55 crc kubenswrapper[4492]: E0220 08:17:55.561043 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:18:06 crc kubenswrapper[4492]: I0220 08:18:06.556762 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:18:06 crc kubenswrapper[4492]: E0220 08:18:06.557593 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:18:21 crc kubenswrapper[4492]: I0220 08:18:21.556904 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:18:21 crc kubenswrapper[4492]: E0220 08:18:21.557778 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:18:32 crc kubenswrapper[4492]: I0220 08:18:32.557923 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:18:32 crc kubenswrapper[4492]: E0220 08:18:32.558803 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:18:45 crc kubenswrapper[4492]: I0220 08:18:45.557601 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:18:45 crc kubenswrapper[4492]: E0220 08:18:45.558445 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:18:57 crc kubenswrapper[4492]: I0220 08:18:57.557147 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:18:57 crc kubenswrapper[4492]: E0220 08:18:57.559140 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:19:12 crc kubenswrapper[4492]: I0220 08:19:12.557607 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:19:12 crc kubenswrapper[4492]: E0220 08:19:12.558657 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:19:25 crc kubenswrapper[4492]: I0220 08:19:25.557716 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:19:25 crc kubenswrapper[4492]: E0220 08:19:25.558534 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:19:37 crc kubenswrapper[4492]: I0220 08:19:37.557298 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:19:37 crc kubenswrapper[4492]: E0220 08:19:37.558321 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:19:52 crc kubenswrapper[4492]: I0220 08:19:52.557207 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:19:53 crc kubenswrapper[4492]: I0220 08:19:53.112390 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"162db159510b88f469186ce670d51ba36bc30287734e7729a38b8bd05ac65d0f"} Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.441652 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v5gw6"] Feb 20 08:21:38 crc kubenswrapper[4492]: E0220 08:21:38.442969 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64892d2a-51dd-4ae6-8936-3c9591b32884" containerName="collect-profiles" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.442992 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="64892d2a-51dd-4ae6-8936-3c9591b32884" containerName="collect-profiles" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.443279 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="64892d2a-51dd-4ae6-8936-3c9591b32884" containerName="collect-profiles" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.445439 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.457512 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5gw6"] Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.503872 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-catalog-content\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.504047 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fng77\" (UniqueName: \"kubernetes.io/projected/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-kube-api-access-fng77\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.504132 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-utilities\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.606653 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-catalog-content\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.606905 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fng77\" (UniqueName: \"kubernetes.io/projected/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-kube-api-access-fng77\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.607002 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-utilities\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.607447 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-catalog-content\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.607867 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-utilities\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.627820 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fng77\" (UniqueName: \"kubernetes.io/projected/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-kube-api-access-fng77\") pod \"certified-operators-v5gw6\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:38 crc kubenswrapper[4492]: I0220 08:21:38.762997 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:39 crc kubenswrapper[4492]: I0220 08:21:39.474557 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v5gw6"] Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.094685 4492 generic.go:334] "Generic (PLEG): container finished" podID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerID="175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018" exitCode=0 Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.095109 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5gw6" event={"ID":"8f9c7a1f-b12b-4327-81d1-a3bb75e46668","Type":"ContainerDied","Data":"175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018"} Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.095155 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5gw6" event={"ID":"8f9c7a1f-b12b-4327-81d1-a3bb75e46668","Type":"ContainerStarted","Data":"87369aee7d63773eeeab1f6f0ce9f78daee70b935f12531096bc84a44bcacba8"} Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.101341 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.844353 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6jcnw"] Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.847802 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.863344 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-catalog-content\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.863452 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-utilities\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.863990 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l8db\" (UniqueName: \"kubernetes.io/projected/6adc7a11-245b-45eb-bd05-fbce395d6340-kube-api-access-4l8db\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.887027 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jcnw"] Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.966528 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l8db\" (UniqueName: \"kubernetes.io/projected/6adc7a11-245b-45eb-bd05-fbce395d6340-kube-api-access-4l8db\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.966711 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-catalog-content\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.966841 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-utilities\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.967403 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-catalog-content\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.970181 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-utilities\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:40 crc kubenswrapper[4492]: I0220 08:21:40.990438 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l8db\" (UniqueName: \"kubernetes.io/projected/6adc7a11-245b-45eb-bd05-fbce395d6340-kube-api-access-4l8db\") pod \"redhat-marketplace-6jcnw\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.197456 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.466126 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sllsn"] Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.470994 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.485073 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sllsn"] Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.586556 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dfh4\" (UniqueName: \"kubernetes.io/projected/ccd9f195-9916-45b5-886b-a793b09b35c0-kube-api-access-7dfh4\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.586692 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-catalog-content\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.586818 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-utilities\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.650240 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jcnw"] Feb 20 08:21:41 crc kubenswrapper[4492]: W0220 08:21:41.657106 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6adc7a11_245b_45eb_bd05_fbce395d6340.slice/crio-a3f30f7320d9243a3d1f690624fc7b9396b9dd67056fa05322afb3a680b1d101 WatchSource:0}: Error finding container a3f30f7320d9243a3d1f690624fc7b9396b9dd67056fa05322afb3a680b1d101: Status 404 returned error can't find the container with id a3f30f7320d9243a3d1f690624fc7b9396b9dd67056fa05322afb3a680b1d101 Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.689212 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dfh4\" (UniqueName: \"kubernetes.io/projected/ccd9f195-9916-45b5-886b-a793b09b35c0-kube-api-access-7dfh4\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.689803 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-catalog-content\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.690496 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-catalog-content\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.690650 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-utilities\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.690966 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-utilities\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.709135 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dfh4\" (UniqueName: \"kubernetes.io/projected/ccd9f195-9916-45b5-886b-a793b09b35c0-kube-api-access-7dfh4\") pod \"redhat-operators-sllsn\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:41 crc kubenswrapper[4492]: I0220 08:21:41.802296 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:42 crc kubenswrapper[4492]: I0220 08:21:42.117880 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5gw6" event={"ID":"8f9c7a1f-b12b-4327-81d1-a3bb75e46668","Type":"ContainerStarted","Data":"b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa"} Feb 20 08:21:42 crc kubenswrapper[4492]: I0220 08:21:42.121809 4492 generic.go:334] "Generic (PLEG): container finished" podID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerID="cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e" exitCode=0 Feb 20 08:21:42 crc kubenswrapper[4492]: I0220 08:21:42.121871 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jcnw" event={"ID":"6adc7a11-245b-45eb-bd05-fbce395d6340","Type":"ContainerDied","Data":"cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e"} Feb 20 08:21:42 crc kubenswrapper[4492]: I0220 08:21:42.121907 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jcnw" event={"ID":"6adc7a11-245b-45eb-bd05-fbce395d6340","Type":"ContainerStarted","Data":"a3f30f7320d9243a3d1f690624fc7b9396b9dd67056fa05322afb3a680b1d101"} Feb 20 08:21:42 crc kubenswrapper[4492]: I0220 08:21:42.371533 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sllsn"] Feb 20 08:21:43 crc kubenswrapper[4492]: I0220 08:21:43.134315 4492 generic.go:334] "Generic (PLEG): container finished" podID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerID="25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e" exitCode=0 Feb 20 08:21:43 crc kubenswrapper[4492]: I0220 08:21:43.135258 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sllsn" event={"ID":"ccd9f195-9916-45b5-886b-a793b09b35c0","Type":"ContainerDied","Data":"25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e"} Feb 20 08:21:43 crc kubenswrapper[4492]: I0220 08:21:43.135317 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sllsn" event={"ID":"ccd9f195-9916-45b5-886b-a793b09b35c0","Type":"ContainerStarted","Data":"955d3d500f45d4ffb5b1ae82c798bf479c7a0b0e3dd56da77b9271d92b6d7b8c"} Feb 20 08:21:43 crc kubenswrapper[4492]: I0220 08:21:43.142546 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jcnw" event={"ID":"6adc7a11-245b-45eb-bd05-fbce395d6340","Type":"ContainerStarted","Data":"58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6"} Feb 20 08:21:43 crc kubenswrapper[4492]: I0220 08:21:43.145616 4492 generic.go:334] "Generic (PLEG): container finished" podID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerID="b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa" exitCode=0 Feb 20 08:21:43 crc kubenswrapper[4492]: I0220 08:21:43.145654 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5gw6" event={"ID":"8f9c7a1f-b12b-4327-81d1-a3bb75e46668","Type":"ContainerDied","Data":"b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa"} Feb 20 08:21:44 crc kubenswrapper[4492]: I0220 08:21:44.157932 4492 generic.go:334] "Generic (PLEG): container finished" podID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerID="58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6" exitCode=0 Feb 20 08:21:44 crc kubenswrapper[4492]: I0220 08:21:44.158082 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jcnw" event={"ID":"6adc7a11-245b-45eb-bd05-fbce395d6340","Type":"ContainerDied","Data":"58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6"} Feb 20 08:21:44 crc kubenswrapper[4492]: I0220 08:21:44.161179 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5gw6" event={"ID":"8f9c7a1f-b12b-4327-81d1-a3bb75e46668","Type":"ContainerStarted","Data":"bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab"} Feb 20 08:21:44 crc kubenswrapper[4492]: I0220 08:21:44.169680 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sllsn" event={"ID":"ccd9f195-9916-45b5-886b-a793b09b35c0","Type":"ContainerStarted","Data":"326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6"} Feb 20 08:21:44 crc kubenswrapper[4492]: I0220 08:21:44.205696 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v5gw6" podStartSLOduration=2.6881606 podStartE2EDuration="6.205675985s" podCreationTimestamp="2026-02-20 08:21:38 +0000 UTC" firstStartedPulling="2026-02-20 08:21:40.099463212 +0000 UTC m=+6056.870752190" lastFinishedPulling="2026-02-20 08:21:43.616978598 +0000 UTC m=+6060.388267575" observedRunningTime="2026-02-20 08:21:44.199215623 +0000 UTC m=+6060.970504601" watchObservedRunningTime="2026-02-20 08:21:44.205675985 +0000 UTC m=+6060.976964963" Feb 20 08:21:45 crc kubenswrapper[4492]: I0220 08:21:45.185705 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jcnw" event={"ID":"6adc7a11-245b-45eb-bd05-fbce395d6340","Type":"ContainerStarted","Data":"a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb"} Feb 20 08:21:45 crc kubenswrapper[4492]: I0220 08:21:45.220865 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6jcnw" podStartSLOduration=2.732310022 podStartE2EDuration="5.220842805s" podCreationTimestamp="2026-02-20 08:21:40 +0000 UTC" firstStartedPulling="2026-02-20 08:21:42.124082327 +0000 UTC m=+6058.895371305" lastFinishedPulling="2026-02-20 08:21:44.61261511 +0000 UTC m=+6061.383904088" observedRunningTime="2026-02-20 08:21:45.219597806 +0000 UTC m=+6061.990886784" watchObservedRunningTime="2026-02-20 08:21:45.220842805 +0000 UTC m=+6061.992131782" Feb 20 08:21:47 crc kubenswrapper[4492]: I0220 08:21:47.208868 4492 generic.go:334] "Generic (PLEG): container finished" podID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerID="326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6" exitCode=0 Feb 20 08:21:47 crc kubenswrapper[4492]: I0220 08:21:47.209033 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sllsn" event={"ID":"ccd9f195-9916-45b5-886b-a793b09b35c0","Type":"ContainerDied","Data":"326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6"} Feb 20 08:21:48 crc kubenswrapper[4492]: I0220 08:21:48.224223 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sllsn" event={"ID":"ccd9f195-9916-45b5-886b-a793b09b35c0","Type":"ContainerStarted","Data":"c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0"} Feb 20 08:21:48 crc kubenswrapper[4492]: I0220 08:21:48.247581 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sllsn" podStartSLOduration=2.670404639 podStartE2EDuration="7.24756294s" podCreationTimestamp="2026-02-20 08:21:41 +0000 UTC" firstStartedPulling="2026-02-20 08:21:43.136680953 +0000 UTC m=+6059.907969931" lastFinishedPulling="2026-02-20 08:21:47.713839254 +0000 UTC m=+6064.485128232" observedRunningTime="2026-02-20 08:21:48.241265976 +0000 UTC m=+6065.012554954" watchObservedRunningTime="2026-02-20 08:21:48.24756294 +0000 UTC m=+6065.018851918" Feb 20 08:21:48 crc kubenswrapper[4492]: I0220 08:21:48.763698 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:48 crc kubenswrapper[4492]: I0220 08:21:48.763758 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:49 crc kubenswrapper[4492]: I0220 08:21:49.809357 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-v5gw6" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="registry-server" probeResult="failure" output=< Feb 20 08:21:49 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:21:49 crc kubenswrapper[4492]: > Feb 20 08:21:51 crc kubenswrapper[4492]: I0220 08:21:51.197770 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:51 crc kubenswrapper[4492]: I0220 08:21:51.197849 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:51 crc kubenswrapper[4492]: I0220 08:21:51.249695 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:51 crc kubenswrapper[4492]: I0220 08:21:51.311184 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:51 crc kubenswrapper[4492]: I0220 08:21:51.802414 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:51 crc kubenswrapper[4492]: I0220 08:21:51.802762 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:21:52 crc kubenswrapper[4492]: I0220 08:21:52.430089 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jcnw"] Feb 20 08:21:52 crc kubenswrapper[4492]: I0220 08:21:52.841418 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sllsn" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="registry-server" probeResult="failure" output=< Feb 20 08:21:52 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:21:52 crc kubenswrapper[4492]: > Feb 20 08:21:53 crc kubenswrapper[4492]: I0220 08:21:53.286626 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6jcnw" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="registry-server" containerID="cri-o://a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb" gracePeriod=2 Feb 20 08:21:53 crc kubenswrapper[4492]: I0220 08:21:53.851685 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.016931 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-utilities\") pod \"6adc7a11-245b-45eb-bd05-fbce395d6340\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.017198 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-catalog-content\") pod \"6adc7a11-245b-45eb-bd05-fbce395d6340\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.017416 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l8db\" (UniqueName: \"kubernetes.io/projected/6adc7a11-245b-45eb-bd05-fbce395d6340-kube-api-access-4l8db\") pod \"6adc7a11-245b-45eb-bd05-fbce395d6340\" (UID: \"6adc7a11-245b-45eb-bd05-fbce395d6340\") " Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.018914 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-utilities" (OuterVolumeSpecName: "utilities") pod "6adc7a11-245b-45eb-bd05-fbce395d6340" (UID: "6adc7a11-245b-45eb-bd05-fbce395d6340"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.026092 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6adc7a11-245b-45eb-bd05-fbce395d6340-kube-api-access-4l8db" (OuterVolumeSpecName: "kube-api-access-4l8db") pod "6adc7a11-245b-45eb-bd05-fbce395d6340" (UID: "6adc7a11-245b-45eb-bd05-fbce395d6340"). InnerVolumeSpecName "kube-api-access-4l8db". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.037122 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6adc7a11-245b-45eb-bd05-fbce395d6340" (UID: "6adc7a11-245b-45eb-bd05-fbce395d6340"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.120202 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l8db\" (UniqueName: \"kubernetes.io/projected/6adc7a11-245b-45eb-bd05-fbce395d6340-kube-api-access-4l8db\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.120242 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.120255 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6adc7a11-245b-45eb-bd05-fbce395d6340-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.298281 4492 generic.go:334] "Generic (PLEG): container finished" podID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerID="a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb" exitCode=0 Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.298355 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jcnw" event={"ID":"6adc7a11-245b-45eb-bd05-fbce395d6340","Type":"ContainerDied","Data":"a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb"} Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.298398 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jcnw" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.298431 4492 scope.go:117] "RemoveContainer" containerID="a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.298407 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jcnw" event={"ID":"6adc7a11-245b-45eb-bd05-fbce395d6340","Type":"ContainerDied","Data":"a3f30f7320d9243a3d1f690624fc7b9396b9dd67056fa05322afb3a680b1d101"} Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.328417 4492 scope.go:117] "RemoveContainer" containerID="58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.336193 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jcnw"] Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.344166 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jcnw"] Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.358104 4492 scope.go:117] "RemoveContainer" containerID="cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.390639 4492 scope.go:117] "RemoveContainer" containerID="a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb" Feb 20 08:21:54 crc kubenswrapper[4492]: E0220 08:21:54.391948 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb\": container with ID starting with a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb not found: ID does not exist" containerID="a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.391982 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb"} err="failed to get container status \"a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb\": rpc error: code = NotFound desc = could not find container \"a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb\": container with ID starting with a4b8b3a15c6f52bd5edae981899e8fc9e7a4a3332bcb70564a5d9e8fb29956cb not found: ID does not exist" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.392004 4492 scope.go:117] "RemoveContainer" containerID="58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6" Feb 20 08:21:54 crc kubenswrapper[4492]: E0220 08:21:54.392243 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6\": container with ID starting with 58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6 not found: ID does not exist" containerID="58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.392270 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6"} err="failed to get container status \"58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6\": rpc error: code = NotFound desc = could not find container \"58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6\": container with ID starting with 58723202a069fa6de6657f01d0d7a8c99aa0e6821527b2f094fbc308053ad4e6 not found: ID does not exist" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.392286 4492 scope.go:117] "RemoveContainer" containerID="cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e" Feb 20 08:21:54 crc kubenswrapper[4492]: E0220 08:21:54.392545 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e\": container with ID starting with cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e not found: ID does not exist" containerID="cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e" Feb 20 08:21:54 crc kubenswrapper[4492]: I0220 08:21:54.392590 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e"} err="failed to get container status \"cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e\": rpc error: code = NotFound desc = could not find container \"cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e\": container with ID starting with cb4971fc219a1d6aba129b93c0ba2d804fc3f041a546aa46bcbc7ab7bd12b63e not found: ID does not exist" Feb 20 08:21:55 crc kubenswrapper[4492]: I0220 08:21:55.568359 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" path="/var/lib/kubelet/pods/6adc7a11-245b-45eb-bd05-fbce395d6340/volumes" Feb 20 08:21:58 crc kubenswrapper[4492]: I0220 08:21:58.799566 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:58 crc kubenswrapper[4492]: I0220 08:21:58.838763 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:21:59 crc kubenswrapper[4492]: I0220 08:21:59.036189 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5gw6"] Feb 20 08:22:00 crc kubenswrapper[4492]: I0220 08:22:00.347349 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v5gw6" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="registry-server" containerID="cri-o://bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab" gracePeriod=2 Feb 20 08:22:00 crc kubenswrapper[4492]: E0220 08:22:00.598109 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f9c7a1f_b12b_4327_81d1_a3bb75e46668.slice/crio-conmon-bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.000415 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.190506 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-catalog-content\") pod \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.190634 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-utilities\") pod \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.191040 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fng77\" (UniqueName: \"kubernetes.io/projected/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-kube-api-access-fng77\") pod \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\" (UID: \"8f9c7a1f-b12b-4327-81d1-a3bb75e46668\") " Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.191825 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-utilities" (OuterVolumeSpecName: "utilities") pod "8f9c7a1f-b12b-4327-81d1-a3bb75e46668" (UID: "8f9c7a1f-b12b-4327-81d1-a3bb75e46668"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.197737 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-kube-api-access-fng77" (OuterVolumeSpecName: "kube-api-access-fng77") pod "8f9c7a1f-b12b-4327-81d1-a3bb75e46668" (UID: "8f9c7a1f-b12b-4327-81d1-a3bb75e46668"). InnerVolumeSpecName "kube-api-access-fng77". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.245756 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f9c7a1f-b12b-4327-81d1-a3bb75e46668" (UID: "8f9c7a1f-b12b-4327-81d1-a3bb75e46668"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.294538 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fng77\" (UniqueName: \"kubernetes.io/projected/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-kube-api-access-fng77\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.294580 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.294597 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f9c7a1f-b12b-4327-81d1-a3bb75e46668-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.360428 4492 generic.go:334] "Generic (PLEG): container finished" podID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerID="bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab" exitCode=0 Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.360518 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5gw6" event={"ID":"8f9c7a1f-b12b-4327-81d1-a3bb75e46668","Type":"ContainerDied","Data":"bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab"} Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.360531 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v5gw6" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.360569 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v5gw6" event={"ID":"8f9c7a1f-b12b-4327-81d1-a3bb75e46668","Type":"ContainerDied","Data":"87369aee7d63773eeeab1f6f0ce9f78daee70b935f12531096bc84a44bcacba8"} Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.360598 4492 scope.go:117] "RemoveContainer" containerID="bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.397274 4492 scope.go:117] "RemoveContainer" containerID="b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.408086 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v5gw6"] Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.416722 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v5gw6"] Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.434975 4492 scope.go:117] "RemoveContainer" containerID="175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.458938 4492 scope.go:117] "RemoveContainer" containerID="bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab" Feb 20 08:22:01 crc kubenswrapper[4492]: E0220 08:22:01.459338 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab\": container with ID starting with bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab not found: ID does not exist" containerID="bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.459376 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab"} err="failed to get container status \"bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab\": rpc error: code = NotFound desc = could not find container \"bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab\": container with ID starting with bc8a0af65c42a4e9bbd8654fb314f0706e3da49396059d49bfe898d0ba86a7ab not found: ID does not exist" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.459404 4492 scope.go:117] "RemoveContainer" containerID="b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa" Feb 20 08:22:01 crc kubenswrapper[4492]: E0220 08:22:01.459898 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa\": container with ID starting with b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa not found: ID does not exist" containerID="b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.459916 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa"} err="failed to get container status \"b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa\": rpc error: code = NotFound desc = could not find container \"b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa\": container with ID starting with b3c549bf4dcd27ff9cee2933ba8bbea9611626f6040566120ab22c81d17bd2fa not found: ID does not exist" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.459928 4492 scope.go:117] "RemoveContainer" containerID="175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018" Feb 20 08:22:01 crc kubenswrapper[4492]: E0220 08:22:01.460432 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018\": container with ID starting with 175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018 not found: ID does not exist" containerID="175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.460452 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018"} err="failed to get container status \"175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018\": rpc error: code = NotFound desc = could not find container \"175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018\": container with ID starting with 175864396c1ce4bb1bbc6adc404d547238a72d8d537405784441321ba1072018 not found: ID does not exist" Feb 20 08:22:01 crc kubenswrapper[4492]: I0220 08:22:01.565582 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" path="/var/lib/kubelet/pods/8f9c7a1f-b12b-4327-81d1-a3bb75e46668/volumes" Feb 20 08:22:02 crc kubenswrapper[4492]: I0220 08:22:02.848326 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sllsn" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="registry-server" probeResult="failure" output=< Feb 20 08:22:02 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:22:02 crc kubenswrapper[4492]: > Feb 20 08:22:09 crc kubenswrapper[4492]: I0220 08:22:09.311164 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:22:09 crc kubenswrapper[4492]: I0220 08:22:09.312315 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:22:11 crc kubenswrapper[4492]: I0220 08:22:11.845396 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:22:11 crc kubenswrapper[4492]: I0220 08:22:11.886871 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:22:12 crc kubenswrapper[4492]: I0220 08:22:12.623166 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sllsn"] Feb 20 08:22:13 crc kubenswrapper[4492]: I0220 08:22:13.460169 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sllsn" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="registry-server" containerID="cri-o://c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0" gracePeriod=2 Feb 20 08:22:13 crc kubenswrapper[4492]: I0220 08:22:13.968194 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.000088 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dfh4\" (UniqueName: \"kubernetes.io/projected/ccd9f195-9916-45b5-886b-a793b09b35c0-kube-api-access-7dfh4\") pod \"ccd9f195-9916-45b5-886b-a793b09b35c0\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.000245 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-catalog-content\") pod \"ccd9f195-9916-45b5-886b-a793b09b35c0\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.000460 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-utilities\") pod \"ccd9f195-9916-45b5-886b-a793b09b35c0\" (UID: \"ccd9f195-9916-45b5-886b-a793b09b35c0\") " Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.002430 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-utilities" (OuterVolumeSpecName: "utilities") pod "ccd9f195-9916-45b5-886b-a793b09b35c0" (UID: "ccd9f195-9916-45b5-886b-a793b09b35c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.019870 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd9f195-9916-45b5-886b-a793b09b35c0-kube-api-access-7dfh4" (OuterVolumeSpecName: "kube-api-access-7dfh4") pod "ccd9f195-9916-45b5-886b-a793b09b35c0" (UID: "ccd9f195-9916-45b5-886b-a793b09b35c0"). InnerVolumeSpecName "kube-api-access-7dfh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.104935 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dfh4\" (UniqueName: \"kubernetes.io/projected/ccd9f195-9916-45b5-886b-a793b09b35c0-kube-api-access-7dfh4\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.104971 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.116791 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccd9f195-9916-45b5-886b-a793b09b35c0" (UID: "ccd9f195-9916-45b5-886b-a793b09b35c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.207169 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccd9f195-9916-45b5-886b-a793b09b35c0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.472867 4492 generic.go:334] "Generic (PLEG): container finished" podID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerID="c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0" exitCode=0 Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.472938 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sllsn" event={"ID":"ccd9f195-9916-45b5-886b-a793b09b35c0","Type":"ContainerDied","Data":"c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0"} Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.472980 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sllsn" event={"ID":"ccd9f195-9916-45b5-886b-a793b09b35c0","Type":"ContainerDied","Data":"955d3d500f45d4ffb5b1ae82c798bf479c7a0b0e3dd56da77b9271d92b6d7b8c"} Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.473001 4492 scope.go:117] "RemoveContainer" containerID="c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.473012 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sllsn" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.504642 4492 scope.go:117] "RemoveContainer" containerID="326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.513636 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sllsn"] Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.527919 4492 scope.go:117] "RemoveContainer" containerID="25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.530521 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sllsn"] Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.564353 4492 scope.go:117] "RemoveContainer" containerID="c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0" Feb 20 08:22:14 crc kubenswrapper[4492]: E0220 08:22:14.564887 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0\": container with ID starting with c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0 not found: ID does not exist" containerID="c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.564925 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0"} err="failed to get container status \"c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0\": rpc error: code = NotFound desc = could not find container \"c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0\": container with ID starting with c37d20e9d77083882b00c24f0073286926a094e46361e3e442e3bfaaeb2bffa0 not found: ID does not exist" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.564950 4492 scope.go:117] "RemoveContainer" containerID="326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6" Feb 20 08:22:14 crc kubenswrapper[4492]: E0220 08:22:14.565286 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6\": container with ID starting with 326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6 not found: ID does not exist" containerID="326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.565314 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6"} err="failed to get container status \"326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6\": rpc error: code = NotFound desc = could not find container \"326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6\": container with ID starting with 326ab26915aaaae499c97f1a547b56e8893344069ee5bf5b264fd5210e87a7c6 not found: ID does not exist" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.565330 4492 scope.go:117] "RemoveContainer" containerID="25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e" Feb 20 08:22:14 crc kubenswrapper[4492]: E0220 08:22:14.565749 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e\": container with ID starting with 25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e not found: ID does not exist" containerID="25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e" Feb 20 08:22:14 crc kubenswrapper[4492]: I0220 08:22:14.565772 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e"} err="failed to get container status \"25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e\": rpc error: code = NotFound desc = could not find container \"25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e\": container with ID starting with 25e91fb4fc09af4aede6d0e77f94761b1cee0011ae201c2d8b6e1687135eb30e not found: ID does not exist" Feb 20 08:22:15 crc kubenswrapper[4492]: I0220 08:22:15.565853 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" path="/var/lib/kubelet/pods/ccd9f195-9916-45b5-886b-a793b09b35c0/volumes" Feb 20 08:22:39 crc kubenswrapper[4492]: I0220 08:22:39.311262 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:22:39 crc kubenswrapper[4492]: I0220 08:22:39.312466 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.311942 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.312620 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.312684 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.313886 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"162db159510b88f469186ce670d51ba36bc30287734e7729a38b8bd05ac65d0f"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.313951 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://162db159510b88f469186ce670d51ba36bc30287734e7729a38b8bd05ac65d0f" gracePeriod=600 Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.980102 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="162db159510b88f469186ce670d51ba36bc30287734e7729a38b8bd05ac65d0f" exitCode=0 Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.980168 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"162db159510b88f469186ce670d51ba36bc30287734e7729a38b8bd05ac65d0f"} Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.980538 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1"} Feb 20 08:23:09 crc kubenswrapper[4492]: I0220 08:23:09.980572 4492 scope.go:117] "RemoveContainer" containerID="65bb0042911637ad89eda878674bd923df5d7b156f6221041063cc47f2bb4449" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.206435 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qrtfb"] Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207792 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="extract-utilities" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207814 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="extract-utilities" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207840 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="extract-content" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207845 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="extract-content" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207857 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="extract-content" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207865 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="extract-content" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207874 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="extract-content" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207879 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="extract-content" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207886 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207907 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207916 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207923 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207944 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207950 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207962 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="extract-utilities" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207967 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="extract-utilities" Feb 20 08:24:30 crc kubenswrapper[4492]: E0220 08:24:30.207979 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="extract-utilities" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.207984 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="extract-utilities" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.208162 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="6adc7a11-245b-45eb-bd05-fbce395d6340" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.208192 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9c7a1f-b12b-4327-81d1-a3bb75e46668" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.208200 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd9f195-9916-45b5-886b-a793b09b35c0" containerName="registry-server" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.209811 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.229100 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qrtfb"] Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.354551 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-utilities\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.354960 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-catalog-content\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.355117 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc957\" (UniqueName: \"kubernetes.io/projected/9af2964a-253e-4c23-9c00-4a1132b6dc63-kube-api-access-lc957\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.458272 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-utilities\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.458754 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-catalog-content\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.458858 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-utilities\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.459007 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc957\" (UniqueName: \"kubernetes.io/projected/9af2964a-253e-4c23-9c00-4a1132b6dc63-kube-api-access-lc957\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.459167 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-catalog-content\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.482759 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc957\" (UniqueName: \"kubernetes.io/projected/9af2964a-253e-4c23-9c00-4a1132b6dc63-kube-api-access-lc957\") pod \"community-operators-qrtfb\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:30 crc kubenswrapper[4492]: I0220 08:24:30.532384 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:31 crc kubenswrapper[4492]: I0220 08:24:31.036329 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qrtfb"] Feb 20 08:24:31 crc kubenswrapper[4492]: I0220 08:24:31.824969 4492 generic.go:334] "Generic (PLEG): container finished" podID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerID="1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0" exitCode=0 Feb 20 08:24:31 crc kubenswrapper[4492]: I0220 08:24:31.825612 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qrtfb" event={"ID":"9af2964a-253e-4c23-9c00-4a1132b6dc63","Type":"ContainerDied","Data":"1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0"} Feb 20 08:24:31 crc kubenswrapper[4492]: I0220 08:24:31.825658 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qrtfb" event={"ID":"9af2964a-253e-4c23-9c00-4a1132b6dc63","Type":"ContainerStarted","Data":"05297d79611fb2f69093206702883af0711776288e7d25fe8ae33c2218ddc948"} Feb 20 08:24:32 crc kubenswrapper[4492]: I0220 08:24:32.865163 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qrtfb" event={"ID":"9af2964a-253e-4c23-9c00-4a1132b6dc63","Type":"ContainerStarted","Data":"176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c"} Feb 20 08:24:33 crc kubenswrapper[4492]: I0220 08:24:33.875289 4492 generic.go:334] "Generic (PLEG): container finished" podID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerID="176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c" exitCode=0 Feb 20 08:24:33 crc kubenswrapper[4492]: I0220 08:24:33.875368 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qrtfb" event={"ID":"9af2964a-253e-4c23-9c00-4a1132b6dc63","Type":"ContainerDied","Data":"176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c"} Feb 20 08:24:34 crc kubenswrapper[4492]: I0220 08:24:34.887454 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qrtfb" event={"ID":"9af2964a-253e-4c23-9c00-4a1132b6dc63","Type":"ContainerStarted","Data":"9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b"} Feb 20 08:24:34 crc kubenswrapper[4492]: I0220 08:24:34.906637 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qrtfb" podStartSLOduration=2.385037241 podStartE2EDuration="4.906612009s" podCreationTimestamp="2026-02-20 08:24:30 +0000 UTC" firstStartedPulling="2026-02-20 08:24:31.829159921 +0000 UTC m=+6228.600448898" lastFinishedPulling="2026-02-20 08:24:34.350734689 +0000 UTC m=+6231.122023666" observedRunningTime="2026-02-20 08:24:34.904377926 +0000 UTC m=+6231.675666905" watchObservedRunningTime="2026-02-20 08:24:34.906612009 +0000 UTC m=+6231.677900987" Feb 20 08:24:40 crc kubenswrapper[4492]: I0220 08:24:40.533394 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:40 crc kubenswrapper[4492]: I0220 08:24:40.534047 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:40 crc kubenswrapper[4492]: I0220 08:24:40.573067 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:40 crc kubenswrapper[4492]: I0220 08:24:40.978586 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:41 crc kubenswrapper[4492]: I0220 08:24:41.026226 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qrtfb"] Feb 20 08:24:42 crc kubenswrapper[4492]: I0220 08:24:42.960419 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qrtfb" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="registry-server" containerID="cri-o://9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b" gracePeriod=2 Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.410509 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.562610 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-utilities\") pod \"9af2964a-253e-4c23-9c00-4a1132b6dc63\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.563911 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-utilities" (OuterVolumeSpecName: "utilities") pod "9af2964a-253e-4c23-9c00-4a1132b6dc63" (UID: "9af2964a-253e-4c23-9c00-4a1132b6dc63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.564406 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-catalog-content\") pod \"9af2964a-253e-4c23-9c00-4a1132b6dc63\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.564561 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc957\" (UniqueName: \"kubernetes.io/projected/9af2964a-253e-4c23-9c00-4a1132b6dc63-kube-api-access-lc957\") pod \"9af2964a-253e-4c23-9c00-4a1132b6dc63\" (UID: \"9af2964a-253e-4c23-9c00-4a1132b6dc63\") " Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.568031 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.572317 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9af2964a-253e-4c23-9c00-4a1132b6dc63-kube-api-access-lc957" (OuterVolumeSpecName: "kube-api-access-lc957") pod "9af2964a-253e-4c23-9c00-4a1132b6dc63" (UID: "9af2964a-253e-4c23-9c00-4a1132b6dc63"). InnerVolumeSpecName "kube-api-access-lc957". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.619730 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9af2964a-253e-4c23-9c00-4a1132b6dc63" (UID: "9af2964a-253e-4c23-9c00-4a1132b6dc63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.673421 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc957\" (UniqueName: \"kubernetes.io/projected/9af2964a-253e-4c23-9c00-4a1132b6dc63-kube-api-access-lc957\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.673453 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9af2964a-253e-4c23-9c00-4a1132b6dc63-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.975682 4492 generic.go:334] "Generic (PLEG): container finished" podID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerID="9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b" exitCode=0 Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.976737 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qrtfb" Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.976865 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qrtfb" event={"ID":"9af2964a-253e-4c23-9c00-4a1132b6dc63","Type":"ContainerDied","Data":"9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b"} Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.977289 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qrtfb" event={"ID":"9af2964a-253e-4c23-9c00-4a1132b6dc63","Type":"ContainerDied","Data":"05297d79611fb2f69093206702883af0711776288e7d25fe8ae33c2218ddc948"} Feb 20 08:24:43 crc kubenswrapper[4492]: I0220 08:24:43.977437 4492 scope.go:117] "RemoveContainer" containerID="9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.019169 4492 scope.go:117] "RemoveContainer" containerID="176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.021298 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qrtfb"] Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.029440 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qrtfb"] Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.039444 4492 scope.go:117] "RemoveContainer" containerID="1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.073841 4492 scope.go:117] "RemoveContainer" containerID="9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b" Feb 20 08:24:44 crc kubenswrapper[4492]: E0220 08:24:44.074196 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b\": container with ID starting with 9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b not found: ID does not exist" containerID="9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.074228 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b"} err="failed to get container status \"9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b\": rpc error: code = NotFound desc = could not find container \"9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b\": container with ID starting with 9bc5caef622088d7229c2866b57cf2c55127081c40315bf3b25fb87c44a7a70b not found: ID does not exist" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.074251 4492 scope.go:117] "RemoveContainer" containerID="176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c" Feb 20 08:24:44 crc kubenswrapper[4492]: E0220 08:24:44.074498 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c\": container with ID starting with 176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c not found: ID does not exist" containerID="176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.074521 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c"} err="failed to get container status \"176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c\": rpc error: code = NotFound desc = could not find container \"176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c\": container with ID starting with 176042f9e95455b2e195004741c540a26d50040c25b4f6dc13ef73ac9677370c not found: ID does not exist" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.074538 4492 scope.go:117] "RemoveContainer" containerID="1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0" Feb 20 08:24:44 crc kubenswrapper[4492]: E0220 08:24:44.074860 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0\": container with ID starting with 1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0 not found: ID does not exist" containerID="1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0" Feb 20 08:24:44 crc kubenswrapper[4492]: I0220 08:24:44.074898 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0"} err="failed to get container status \"1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0\": rpc error: code = NotFound desc = could not find container \"1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0\": container with ID starting with 1b10a7d3914ba95d1874b3787aa33c463cecd12116c7aaeb0c6db4d1d85cd0c0 not found: ID does not exist" Feb 20 08:24:45 crc kubenswrapper[4492]: I0220 08:24:45.567395 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" path="/var/lib/kubelet/pods/9af2964a-253e-4c23-9c00-4a1132b6dc63/volumes" Feb 20 08:25:09 crc kubenswrapper[4492]: I0220 08:25:09.310998 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:25:09 crc kubenswrapper[4492]: I0220 08:25:09.311773 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:25:39 crc kubenswrapper[4492]: I0220 08:25:39.311877 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:25:39 crc kubenswrapper[4492]: I0220 08:25:39.312680 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.311432 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.312354 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.312440 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.313436 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.313520 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" gracePeriod=600 Feb 20 08:26:09 crc kubenswrapper[4492]: E0220 08:26:09.431319 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.792862 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" exitCode=0 Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.793068 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1"} Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.793278 4492 scope.go:117] "RemoveContainer" containerID="162db159510b88f469186ce670d51ba36bc30287734e7729a38b8bd05ac65d0f" Feb 20 08:26:09 crc kubenswrapper[4492]: I0220 08:26:09.794198 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:26:09 crc kubenswrapper[4492]: E0220 08:26:09.794606 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:26:20 crc kubenswrapper[4492]: I0220 08:26:20.556981 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:26:20 crc kubenswrapper[4492]: E0220 08:26:20.558029 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:26:34 crc kubenswrapper[4492]: I0220 08:26:34.561849 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:26:34 crc kubenswrapper[4492]: E0220 08:26:34.563338 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:26:48 crc kubenswrapper[4492]: I0220 08:26:48.557089 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:26:48 crc kubenswrapper[4492]: E0220 08:26:48.558184 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:27:02 crc kubenswrapper[4492]: I0220 08:27:02.556672 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:27:02 crc kubenswrapper[4492]: E0220 08:27:02.558835 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:27:15 crc kubenswrapper[4492]: I0220 08:27:15.556994 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:27:15 crc kubenswrapper[4492]: E0220 08:27:15.557991 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:27:29 crc kubenswrapper[4492]: I0220 08:27:29.558124 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:27:29 crc kubenswrapper[4492]: E0220 08:27:29.559208 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:27:42 crc kubenswrapper[4492]: I0220 08:27:42.558295 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:27:42 crc kubenswrapper[4492]: E0220 08:27:42.559418 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:27:57 crc kubenswrapper[4492]: I0220 08:27:57.558975 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:27:57 crc kubenswrapper[4492]: E0220 08:27:57.560192 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:28:12 crc kubenswrapper[4492]: I0220 08:28:12.557060 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:28:12 crc kubenswrapper[4492]: E0220 08:28:12.558111 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:28:24 crc kubenswrapper[4492]: I0220 08:28:24.557755 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:28:24 crc kubenswrapper[4492]: E0220 08:28:24.558737 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:28:39 crc kubenswrapper[4492]: I0220 08:28:39.557387 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:28:39 crc kubenswrapper[4492]: E0220 08:28:39.558168 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:28:50 crc kubenswrapper[4492]: I0220 08:28:50.558342 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:28:50 crc kubenswrapper[4492]: E0220 08:28:50.559338 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:29:04 crc kubenswrapper[4492]: I0220 08:29:04.558269 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:29:04 crc kubenswrapper[4492]: E0220 08:29:04.559692 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:29:16 crc kubenswrapper[4492]: I0220 08:29:16.557437 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:29:16 crc kubenswrapper[4492]: E0220 08:29:16.558545 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:29:28 crc kubenswrapper[4492]: I0220 08:29:28.557345 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:29:28 crc kubenswrapper[4492]: E0220 08:29:28.558322 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:29:41 crc kubenswrapper[4492]: I0220 08:29:41.558282 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:29:41 crc kubenswrapper[4492]: E0220 08:29:41.558983 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:29:53 crc kubenswrapper[4492]: I0220 08:29:53.563746 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:29:53 crc kubenswrapper[4492]: E0220 08:29:53.565948 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.152185 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh"] Feb 20 08:30:00 crc kubenswrapper[4492]: E0220 08:30:00.153097 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="extract-content" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.153112 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="extract-content" Feb 20 08:30:00 crc kubenswrapper[4492]: E0220 08:30:00.153123 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="registry-server" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.153129 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="registry-server" Feb 20 08:30:00 crc kubenswrapper[4492]: E0220 08:30:00.153157 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="extract-utilities" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.153162 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="extract-utilities" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.153342 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af2964a-253e-4c23-9c00-4a1132b6dc63" containerName="registry-server" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.153997 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.162417 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.167999 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh"] Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.179318 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.191974 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae3b0337-8c14-4f19-9c88-6da340069e0d-secret-volume\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.192138 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4qx7\" (UniqueName: \"kubernetes.io/projected/ae3b0337-8c14-4f19-9c88-6da340069e0d-kube-api-access-r4qx7\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.192168 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae3b0337-8c14-4f19-9c88-6da340069e0d-config-volume\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.294551 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4qx7\" (UniqueName: \"kubernetes.io/projected/ae3b0337-8c14-4f19-9c88-6da340069e0d-kube-api-access-r4qx7\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.294609 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae3b0337-8c14-4f19-9c88-6da340069e0d-config-volume\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.294862 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae3b0337-8c14-4f19-9c88-6da340069e0d-secret-volume\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.296039 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae3b0337-8c14-4f19-9c88-6da340069e0d-config-volume\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.306852 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae3b0337-8c14-4f19-9c88-6da340069e0d-secret-volume\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.310874 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4qx7\" (UniqueName: \"kubernetes.io/projected/ae3b0337-8c14-4f19-9c88-6da340069e0d-kube-api-access-r4qx7\") pod \"collect-profiles-29526270-clhlh\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:00 crc kubenswrapper[4492]: I0220 08:30:00.472784 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:01 crc kubenswrapper[4492]: I0220 08:30:01.035758 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh"] Feb 20 08:30:01 crc kubenswrapper[4492]: I0220 08:30:01.087980 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" event={"ID":"ae3b0337-8c14-4f19-9c88-6da340069e0d","Type":"ContainerStarted","Data":"1445bd8213599b5ac1435e00f4568182cf34a1b78ba93ccb16779c5726e36d18"} Feb 20 08:30:02 crc kubenswrapper[4492]: I0220 08:30:02.099362 4492 generic.go:334] "Generic (PLEG): container finished" podID="ae3b0337-8c14-4f19-9c88-6da340069e0d" containerID="ff8c07c9b1493eb4cb24e4a14e42f263f201df801d495b6fcc5c57d09a9caaac" exitCode=0 Feb 20 08:30:02 crc kubenswrapper[4492]: I0220 08:30:02.099465 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" event={"ID":"ae3b0337-8c14-4f19-9c88-6da340069e0d","Type":"ContainerDied","Data":"ff8c07c9b1493eb4cb24e4a14e42f263f201df801d495b6fcc5c57d09a9caaac"} Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.438091 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.481382 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae3b0337-8c14-4f19-9c88-6da340069e0d-secret-volume\") pod \"ae3b0337-8c14-4f19-9c88-6da340069e0d\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.481445 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4qx7\" (UniqueName: \"kubernetes.io/projected/ae3b0337-8c14-4f19-9c88-6da340069e0d-kube-api-access-r4qx7\") pod \"ae3b0337-8c14-4f19-9c88-6da340069e0d\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.481676 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae3b0337-8c14-4f19-9c88-6da340069e0d-config-volume\") pod \"ae3b0337-8c14-4f19-9c88-6da340069e0d\" (UID: \"ae3b0337-8c14-4f19-9c88-6da340069e0d\") " Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.483605 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae3b0337-8c14-4f19-9c88-6da340069e0d-config-volume" (OuterVolumeSpecName: "config-volume") pod "ae3b0337-8c14-4f19-9c88-6da340069e0d" (UID: "ae3b0337-8c14-4f19-9c88-6da340069e0d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.493129 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3b0337-8c14-4f19-9c88-6da340069e0d-kube-api-access-r4qx7" (OuterVolumeSpecName: "kube-api-access-r4qx7") pod "ae3b0337-8c14-4f19-9c88-6da340069e0d" (UID: "ae3b0337-8c14-4f19-9c88-6da340069e0d"). InnerVolumeSpecName "kube-api-access-r4qx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.493821 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3b0337-8c14-4f19-9c88-6da340069e0d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ae3b0337-8c14-4f19-9c88-6da340069e0d" (UID: "ae3b0337-8c14-4f19-9c88-6da340069e0d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.584067 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ae3b0337-8c14-4f19-9c88-6da340069e0d-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.584100 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ae3b0337-8c14-4f19-9c88-6da340069e0d-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:03 crc kubenswrapper[4492]: I0220 08:30:03.584112 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4qx7\" (UniqueName: \"kubernetes.io/projected/ae3b0337-8c14-4f19-9c88-6da340069e0d-kube-api-access-r4qx7\") on node \"crc\" DevicePath \"\"" Feb 20 08:30:04 crc kubenswrapper[4492]: I0220 08:30:04.118913 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" event={"ID":"ae3b0337-8c14-4f19-9c88-6da340069e0d","Type":"ContainerDied","Data":"1445bd8213599b5ac1435e00f4568182cf34a1b78ba93ccb16779c5726e36d18"} Feb 20 08:30:04 crc kubenswrapper[4492]: I0220 08:30:04.118995 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh" Feb 20 08:30:04 crc kubenswrapper[4492]: I0220 08:30:04.119286 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1445bd8213599b5ac1435e00f4568182cf34a1b78ba93ccb16779c5726e36d18" Feb 20 08:30:04 crc kubenswrapper[4492]: I0220 08:30:04.529869 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp"] Feb 20 08:30:04 crc kubenswrapper[4492]: I0220 08:30:04.535132 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526225-5zqhp"] Feb 20 08:30:05 crc kubenswrapper[4492]: I0220 08:30:05.573733 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80bab9bc-ffef-40d6-8a61-a834994983d5" path="/var/lib/kubelet/pods/80bab9bc-ffef-40d6-8a61-a834994983d5/volumes" Feb 20 08:30:07 crc kubenswrapper[4492]: I0220 08:30:07.557889 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:30:07 crc kubenswrapper[4492]: E0220 08:30:07.558379 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:30:09 crc kubenswrapper[4492]: I0220 08:30:09.752455 4492 scope.go:117] "RemoveContainer" containerID="a129bb0d8c8469d6031580ca1e3f7b1ad9a4ae0e0f35bb28288082d423ead4fd" Feb 20 08:30:22 crc kubenswrapper[4492]: I0220 08:30:22.556877 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:30:22 crc kubenswrapper[4492]: E0220 08:30:22.557947 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:30:36 crc kubenswrapper[4492]: I0220 08:30:36.557544 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:30:36 crc kubenswrapper[4492]: E0220 08:30:36.558467 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:30:48 crc kubenswrapper[4492]: I0220 08:30:48.557552 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:30:48 crc kubenswrapper[4492]: E0220 08:30:48.558541 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:30:59 crc kubenswrapper[4492]: I0220 08:30:59.557161 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:30:59 crc kubenswrapper[4492]: E0220 08:30:59.558120 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:31:11 crc kubenswrapper[4492]: I0220 08:31:11.558036 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:31:12 crc kubenswrapper[4492]: I0220 08:31:12.745344 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"1e98f704439beef93eb8425535fe35b539ac237fdf8b7df91e9e87c9064a354a"} Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.774128 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-26rf2"] Feb 20 08:31:46 crc kubenswrapper[4492]: E0220 08:31:46.775166 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3b0337-8c14-4f19-9c88-6da340069e0d" containerName="collect-profiles" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.775183 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3b0337-8c14-4f19-9c88-6da340069e0d" containerName="collect-profiles" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.775452 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3b0337-8c14-4f19-9c88-6da340069e0d" containerName="collect-profiles" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.778673 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.797933 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-26rf2"] Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.809648 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxgr9\" (UniqueName: \"kubernetes.io/projected/4f9597a9-7847-4e72-bdb4-df1380f44f6a-kube-api-access-gxgr9\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.809694 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-catalog-content\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.809840 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-utilities\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.911436 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxgr9\" (UniqueName: \"kubernetes.io/projected/4f9597a9-7847-4e72-bdb4-df1380f44f6a-kube-api-access-gxgr9\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.911511 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-catalog-content\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.911572 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-utilities\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.912025 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-utilities\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.912264 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-catalog-content\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:46 crc kubenswrapper[4492]: I0220 08:31:46.937163 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxgr9\" (UniqueName: \"kubernetes.io/projected/4f9597a9-7847-4e72-bdb4-df1380f44f6a-kube-api-access-gxgr9\") pod \"redhat-marketplace-26rf2\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:47 crc kubenswrapper[4492]: I0220 08:31:47.101803 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:47 crc kubenswrapper[4492]: I0220 08:31:47.620281 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-26rf2"] Feb 20 08:31:48 crc kubenswrapper[4492]: I0220 08:31:48.092201 4492 generic.go:334] "Generic (PLEG): container finished" podID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerID="0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37" exitCode=0 Feb 20 08:31:48 crc kubenswrapper[4492]: I0220 08:31:48.092655 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26rf2" event={"ID":"4f9597a9-7847-4e72-bdb4-df1380f44f6a","Type":"ContainerDied","Data":"0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37"} Feb 20 08:31:48 crc kubenswrapper[4492]: I0220 08:31:48.092719 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26rf2" event={"ID":"4f9597a9-7847-4e72-bdb4-df1380f44f6a","Type":"ContainerStarted","Data":"2f49c9772457cda60435f125c24b26291a8e64c84af01388c6318a772e72656a"} Feb 20 08:31:48 crc kubenswrapper[4492]: I0220 08:31:48.101025 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:31:49 crc kubenswrapper[4492]: I0220 08:31:49.102400 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26rf2" event={"ID":"4f9597a9-7847-4e72-bdb4-df1380f44f6a","Type":"ContainerStarted","Data":"86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b"} Feb 20 08:31:50 crc kubenswrapper[4492]: I0220 08:31:50.115683 4492 generic.go:334] "Generic (PLEG): container finished" podID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerID="86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b" exitCode=0 Feb 20 08:31:50 crc kubenswrapper[4492]: I0220 08:31:50.115794 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26rf2" event={"ID":"4f9597a9-7847-4e72-bdb4-df1380f44f6a","Type":"ContainerDied","Data":"86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b"} Feb 20 08:31:51 crc kubenswrapper[4492]: I0220 08:31:51.132277 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26rf2" event={"ID":"4f9597a9-7847-4e72-bdb4-df1380f44f6a","Type":"ContainerStarted","Data":"204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041"} Feb 20 08:31:51 crc kubenswrapper[4492]: I0220 08:31:51.157240 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-26rf2" podStartSLOduration=2.684813981 podStartE2EDuration="5.156519635s" podCreationTimestamp="2026-02-20 08:31:46 +0000 UTC" firstStartedPulling="2026-02-20 08:31:48.09710071 +0000 UTC m=+6664.868389688" lastFinishedPulling="2026-02-20 08:31:50.568806364 +0000 UTC m=+6667.340095342" observedRunningTime="2026-02-20 08:31:51.152267537 +0000 UTC m=+6667.923556515" watchObservedRunningTime="2026-02-20 08:31:51.156519635 +0000 UTC m=+6667.927808613" Feb 20 08:31:57 crc kubenswrapper[4492]: I0220 08:31:57.102988 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:57 crc kubenswrapper[4492]: I0220 08:31:57.103950 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:57 crc kubenswrapper[4492]: I0220 08:31:57.150950 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:57 crc kubenswrapper[4492]: I0220 08:31:57.249102 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:57 crc kubenswrapper[4492]: I0220 08:31:57.393920 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-26rf2"] Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.216875 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-26rf2" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="registry-server" containerID="cri-o://204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041" gracePeriod=2 Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.720622 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.809991 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-utilities\") pod \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.810207 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-catalog-content\") pod \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.810882 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-utilities" (OuterVolumeSpecName: "utilities") pod "4f9597a9-7847-4e72-bdb4-df1380f44f6a" (UID: "4f9597a9-7847-4e72-bdb4-df1380f44f6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.811094 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.828945 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4f9597a9-7847-4e72-bdb4-df1380f44f6a" (UID: "4f9597a9-7847-4e72-bdb4-df1380f44f6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.912329 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxgr9\" (UniqueName: \"kubernetes.io/projected/4f9597a9-7847-4e72-bdb4-df1380f44f6a-kube-api-access-gxgr9\") pod \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\" (UID: \"4f9597a9-7847-4e72-bdb4-df1380f44f6a\") " Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.912774 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f9597a9-7847-4e72-bdb4-df1380f44f6a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:31:59 crc kubenswrapper[4492]: I0220 08:31:59.923281 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f9597a9-7847-4e72-bdb4-df1380f44f6a-kube-api-access-gxgr9" (OuterVolumeSpecName: "kube-api-access-gxgr9") pod "4f9597a9-7847-4e72-bdb4-df1380f44f6a" (UID: "4f9597a9-7847-4e72-bdb4-df1380f44f6a"). InnerVolumeSpecName "kube-api-access-gxgr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.014821 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxgr9\" (UniqueName: \"kubernetes.io/projected/4f9597a9-7847-4e72-bdb4-df1380f44f6a-kube-api-access-gxgr9\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.231921 4492 generic.go:334] "Generic (PLEG): container finished" podID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerID="204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041" exitCode=0 Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.231987 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26rf2" event={"ID":"4f9597a9-7847-4e72-bdb4-df1380f44f6a","Type":"ContainerDied","Data":"204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041"} Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.232028 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-26rf2" event={"ID":"4f9597a9-7847-4e72-bdb4-df1380f44f6a","Type":"ContainerDied","Data":"2f49c9772457cda60435f125c24b26291a8e64c84af01388c6318a772e72656a"} Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.232067 4492 scope.go:117] "RemoveContainer" containerID="204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.232260 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-26rf2" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.261632 4492 scope.go:117] "RemoveContainer" containerID="86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.269111 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-26rf2"] Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.283537 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-26rf2"] Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.296481 4492 scope.go:117] "RemoveContainer" containerID="0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.322744 4492 scope.go:117] "RemoveContainer" containerID="204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041" Feb 20 08:32:00 crc kubenswrapper[4492]: E0220 08:32:00.323256 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041\": container with ID starting with 204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041 not found: ID does not exist" containerID="204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.323289 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041"} err="failed to get container status \"204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041\": rpc error: code = NotFound desc = could not find container \"204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041\": container with ID starting with 204bb2816f2ede61ca5953e437b4cfbf340f95370154b1238b59837c6fed6041 not found: ID does not exist" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.323311 4492 scope.go:117] "RemoveContainer" containerID="86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b" Feb 20 08:32:00 crc kubenswrapper[4492]: E0220 08:32:00.323794 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b\": container with ID starting with 86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b not found: ID does not exist" containerID="86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.323813 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b"} err="failed to get container status \"86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b\": rpc error: code = NotFound desc = could not find container \"86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b\": container with ID starting with 86b113a9db09352c9b90bc80c27c90e84ad5aaa21f07691e0686daf9938d247b not found: ID does not exist" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.323826 4492 scope.go:117] "RemoveContainer" containerID="0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37" Feb 20 08:32:00 crc kubenswrapper[4492]: E0220 08:32:00.324087 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37\": container with ID starting with 0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37 not found: ID does not exist" containerID="0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37" Feb 20 08:32:00 crc kubenswrapper[4492]: I0220 08:32:00.324103 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37"} err="failed to get container status \"0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37\": rpc error: code = NotFound desc = could not find container \"0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37\": container with ID starting with 0fd5ac9d96893e5caa06feed3c2dcf1f6f1f8cfda726697ad404bf446eb2dd37 not found: ID does not exist" Feb 20 08:32:01 crc kubenswrapper[4492]: I0220 08:32:01.568209 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" path="/var/lib/kubelet/pods/4f9597a9-7847-4e72-bdb4-df1380f44f6a/volumes" Feb 20 08:32:14 crc kubenswrapper[4492]: I0220 08:32:14.982057 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-trvjx"] Feb 20 08:32:14 crc kubenswrapper[4492]: E0220 08:32:14.983009 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="registry-server" Feb 20 08:32:14 crc kubenswrapper[4492]: I0220 08:32:14.983024 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="registry-server" Feb 20 08:32:14 crc kubenswrapper[4492]: E0220 08:32:14.983045 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="extract-content" Feb 20 08:32:14 crc kubenswrapper[4492]: I0220 08:32:14.983051 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="extract-content" Feb 20 08:32:14 crc kubenswrapper[4492]: E0220 08:32:14.983063 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="extract-utilities" Feb 20 08:32:14 crc kubenswrapper[4492]: I0220 08:32:14.983068 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="extract-utilities" Feb 20 08:32:14 crc kubenswrapper[4492]: I0220 08:32:14.983255 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f9597a9-7847-4e72-bdb4-df1380f44f6a" containerName="registry-server" Feb 20 08:32:14 crc kubenswrapper[4492]: I0220 08:32:14.984456 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:14 crc kubenswrapper[4492]: I0220 08:32:14.999939 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-trvjx"] Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.057115 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-utilities\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.057302 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-catalog-content\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.057402 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhdtl\" (UniqueName: \"kubernetes.io/projected/40d62045-770a-4587-a375-0ba39b046e70-kube-api-access-jhdtl\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.159053 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-catalog-content\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.159127 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhdtl\" (UniqueName: \"kubernetes.io/projected/40d62045-770a-4587-a375-0ba39b046e70-kube-api-access-jhdtl\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.159188 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-utilities\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.159542 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-catalog-content\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.159768 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-utilities\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.184401 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhdtl\" (UniqueName: \"kubernetes.io/projected/40d62045-770a-4587-a375-0ba39b046e70-kube-api-access-jhdtl\") pod \"redhat-operators-trvjx\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:15 crc kubenswrapper[4492]: I0220 08:32:15.305901 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:16 crc kubenswrapper[4492]: I0220 08:32:16.349979 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-trvjx"] Feb 20 08:32:16 crc kubenswrapper[4492]: I0220 08:32:16.404848 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trvjx" event={"ID":"40d62045-770a-4587-a375-0ba39b046e70","Type":"ContainerStarted","Data":"22a18f8b7ebb3d08df6a0e6140c6da13d3bfc5a57cd9dc62affa745b353e7455"} Feb 20 08:32:17 crc kubenswrapper[4492]: I0220 08:32:17.415011 4492 generic.go:334] "Generic (PLEG): container finished" podID="40d62045-770a-4587-a375-0ba39b046e70" containerID="fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f" exitCode=0 Feb 20 08:32:17 crc kubenswrapper[4492]: I0220 08:32:17.415322 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trvjx" event={"ID":"40d62045-770a-4587-a375-0ba39b046e70","Type":"ContainerDied","Data":"fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f"} Feb 20 08:32:18 crc kubenswrapper[4492]: I0220 08:32:18.424640 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trvjx" event={"ID":"40d62045-770a-4587-a375-0ba39b046e70","Type":"ContainerStarted","Data":"01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a"} Feb 20 08:32:21 crc kubenswrapper[4492]: I0220 08:32:21.450516 4492 generic.go:334] "Generic (PLEG): container finished" podID="40d62045-770a-4587-a375-0ba39b046e70" containerID="01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a" exitCode=0 Feb 20 08:32:21 crc kubenswrapper[4492]: I0220 08:32:21.450589 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trvjx" event={"ID":"40d62045-770a-4587-a375-0ba39b046e70","Type":"ContainerDied","Data":"01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a"} Feb 20 08:32:22 crc kubenswrapper[4492]: I0220 08:32:22.464492 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trvjx" event={"ID":"40d62045-770a-4587-a375-0ba39b046e70","Type":"ContainerStarted","Data":"e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01"} Feb 20 08:32:22 crc kubenswrapper[4492]: I0220 08:32:22.489889 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-trvjx" podStartSLOduration=3.983105669 podStartE2EDuration="8.489870801s" podCreationTimestamp="2026-02-20 08:32:14 +0000 UTC" firstStartedPulling="2026-02-20 08:32:17.419313744 +0000 UTC m=+6694.190602722" lastFinishedPulling="2026-02-20 08:32:21.926078877 +0000 UTC m=+6698.697367854" observedRunningTime="2026-02-20 08:32:22.484256685 +0000 UTC m=+6699.255545663" watchObservedRunningTime="2026-02-20 08:32:22.489870801 +0000 UTC m=+6699.261159780" Feb 20 08:32:25 crc kubenswrapper[4492]: I0220 08:32:25.306971 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:25 crc kubenswrapper[4492]: I0220 08:32:25.307767 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:26 crc kubenswrapper[4492]: I0220 08:32:26.351670 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-trvjx" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="registry-server" probeResult="failure" output=< Feb 20 08:32:26 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:32:26 crc kubenswrapper[4492]: > Feb 20 08:32:35 crc kubenswrapper[4492]: I0220 08:32:35.352794 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:35 crc kubenswrapper[4492]: I0220 08:32:35.405752 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:35 crc kubenswrapper[4492]: I0220 08:32:35.600055 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-trvjx"] Feb 20 08:32:36 crc kubenswrapper[4492]: I0220 08:32:36.622556 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-trvjx" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="registry-server" containerID="cri-o://e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01" gracePeriod=2 Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.200865 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.376545 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-catalog-content\") pod \"40d62045-770a-4587-a375-0ba39b046e70\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.376821 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhdtl\" (UniqueName: \"kubernetes.io/projected/40d62045-770a-4587-a375-0ba39b046e70-kube-api-access-jhdtl\") pod \"40d62045-770a-4587-a375-0ba39b046e70\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.377055 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-utilities\") pod \"40d62045-770a-4587-a375-0ba39b046e70\" (UID: \"40d62045-770a-4587-a375-0ba39b046e70\") " Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.378448 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-utilities" (OuterVolumeSpecName: "utilities") pod "40d62045-770a-4587-a375-0ba39b046e70" (UID: "40d62045-770a-4587-a375-0ba39b046e70"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.386027 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40d62045-770a-4587-a375-0ba39b046e70-kube-api-access-jhdtl" (OuterVolumeSpecName: "kube-api-access-jhdtl") pod "40d62045-770a-4587-a375-0ba39b046e70" (UID: "40d62045-770a-4587-a375-0ba39b046e70"). InnerVolumeSpecName "kube-api-access-jhdtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.482149 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhdtl\" (UniqueName: \"kubernetes.io/projected/40d62045-770a-4587-a375-0ba39b046e70-kube-api-access-jhdtl\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.482190 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.495026 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40d62045-770a-4587-a375-0ba39b046e70" (UID: "40d62045-770a-4587-a375-0ba39b046e70"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.584812 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d62045-770a-4587-a375-0ba39b046e70-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.635453 4492 generic.go:334] "Generic (PLEG): container finished" podID="40d62045-770a-4587-a375-0ba39b046e70" containerID="e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01" exitCode=0 Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.635568 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-trvjx" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.636405 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trvjx" event={"ID":"40d62045-770a-4587-a375-0ba39b046e70","Type":"ContainerDied","Data":"e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01"} Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.636675 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-trvjx" event={"ID":"40d62045-770a-4587-a375-0ba39b046e70","Type":"ContainerDied","Data":"22a18f8b7ebb3d08df6a0e6140c6da13d3bfc5a57cd9dc62affa745b353e7455"} Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.636769 4492 scope.go:117] "RemoveContainer" containerID="e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.663824 4492 scope.go:117] "RemoveContainer" containerID="01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.667259 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-trvjx"] Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.678587 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-trvjx"] Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.712104 4492 scope.go:117] "RemoveContainer" containerID="fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.730045 4492 scope.go:117] "RemoveContainer" containerID="e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01" Feb 20 08:32:37 crc kubenswrapper[4492]: E0220 08:32:37.730371 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01\": container with ID starting with e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01 not found: ID does not exist" containerID="e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.730417 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01"} err="failed to get container status \"e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01\": rpc error: code = NotFound desc = could not find container \"e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01\": container with ID starting with e9733b1c003e19ddc3a29945386a822b8e7fdef0ff4086ca4ae9176e2f4dae01 not found: ID does not exist" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.730450 4492 scope.go:117] "RemoveContainer" containerID="01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a" Feb 20 08:32:37 crc kubenswrapper[4492]: E0220 08:32:37.730793 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a\": container with ID starting with 01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a not found: ID does not exist" containerID="01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.730825 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a"} err="failed to get container status \"01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a\": rpc error: code = NotFound desc = could not find container \"01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a\": container with ID starting with 01f7e30d08a2aa71f5de29bbf1ad5d0e4bf3a1ab78629d737ff13ea46fb5b18a not found: ID does not exist" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.730848 4492 scope.go:117] "RemoveContainer" containerID="fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f" Feb 20 08:32:37 crc kubenswrapper[4492]: E0220 08:32:37.731257 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f\": container with ID starting with fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f not found: ID does not exist" containerID="fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f" Feb 20 08:32:37 crc kubenswrapper[4492]: I0220 08:32:37.731293 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f"} err="failed to get container status \"fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f\": rpc error: code = NotFound desc = could not find container \"fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f\": container with ID starting with fd54e545f76a85d40657150b2b2959a60571c83370e01e8b48c50717b705b50f not found: ID does not exist" Feb 20 08:32:39 crc kubenswrapper[4492]: I0220 08:32:39.569576 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40d62045-770a-4587-a375-0ba39b046e70" path="/var/lib/kubelet/pods/40d62045-770a-4587-a375-0ba39b046e70/volumes" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.807341 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mp4xc"] Feb 20 08:32:49 crc kubenswrapper[4492]: E0220 08:32:49.808670 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="extract-utilities" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.808690 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="extract-utilities" Feb 20 08:32:49 crc kubenswrapper[4492]: E0220 08:32:49.808717 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="extract-content" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.808727 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="extract-content" Feb 20 08:32:49 crc kubenswrapper[4492]: E0220 08:32:49.808773 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="registry-server" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.808782 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="registry-server" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.809095 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="40d62045-770a-4587-a375-0ba39b046e70" containerName="registry-server" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.810676 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.839550 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mp4xc"] Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.875144 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2h4x\" (UniqueName: \"kubernetes.io/projected/a82bdc05-86e1-4ae6-a968-30852271a2dc-kube-api-access-j2h4x\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.875435 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-catalog-content\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.875668 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-utilities\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.977703 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-catalog-content\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.977906 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-utilities\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.978046 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2h4x\" (UniqueName: \"kubernetes.io/projected/a82bdc05-86e1-4ae6-a968-30852271a2dc-kube-api-access-j2h4x\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.978337 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-catalog-content\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:49 crc kubenswrapper[4492]: I0220 08:32:49.978393 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-utilities\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:50 crc kubenswrapper[4492]: I0220 08:32:49.998300 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2h4x\" (UniqueName: \"kubernetes.io/projected/a82bdc05-86e1-4ae6-a968-30852271a2dc-kube-api-access-j2h4x\") pod \"certified-operators-mp4xc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:50 crc kubenswrapper[4492]: I0220 08:32:50.142469 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:32:50 crc kubenswrapper[4492]: I0220 08:32:50.623208 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mp4xc"] Feb 20 08:32:50 crc kubenswrapper[4492]: I0220 08:32:50.766620 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp4xc" event={"ID":"a82bdc05-86e1-4ae6-a968-30852271a2dc","Type":"ContainerStarted","Data":"33526abd88a84d6472950584b2dc8faf2bb21615dae17f2aedcbb7cc991935fe"} Feb 20 08:32:51 crc kubenswrapper[4492]: I0220 08:32:51.778653 4492 generic.go:334] "Generic (PLEG): container finished" podID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerID="603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c" exitCode=0 Feb 20 08:32:51 crc kubenswrapper[4492]: I0220 08:32:51.778757 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp4xc" event={"ID":"a82bdc05-86e1-4ae6-a968-30852271a2dc","Type":"ContainerDied","Data":"603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c"} Feb 20 08:32:52 crc kubenswrapper[4492]: I0220 08:32:52.790059 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp4xc" event={"ID":"a82bdc05-86e1-4ae6-a968-30852271a2dc","Type":"ContainerStarted","Data":"e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5"} Feb 20 08:32:54 crc kubenswrapper[4492]: I0220 08:32:54.810284 4492 generic.go:334] "Generic (PLEG): container finished" podID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerID="e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5" exitCode=0 Feb 20 08:32:54 crc kubenswrapper[4492]: I0220 08:32:54.810364 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp4xc" event={"ID":"a82bdc05-86e1-4ae6-a968-30852271a2dc","Type":"ContainerDied","Data":"e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5"} Feb 20 08:32:55 crc kubenswrapper[4492]: I0220 08:32:55.830976 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp4xc" event={"ID":"a82bdc05-86e1-4ae6-a968-30852271a2dc","Type":"ContainerStarted","Data":"0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79"} Feb 20 08:32:55 crc kubenswrapper[4492]: I0220 08:32:55.857311 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mp4xc" podStartSLOduration=3.372433646 podStartE2EDuration="6.857293794s" podCreationTimestamp="2026-02-20 08:32:49 +0000 UTC" firstStartedPulling="2026-02-20 08:32:51.782576021 +0000 UTC m=+6728.553864999" lastFinishedPulling="2026-02-20 08:32:55.267436169 +0000 UTC m=+6732.038725147" observedRunningTime="2026-02-20 08:32:55.855246032 +0000 UTC m=+6732.626535011" watchObservedRunningTime="2026-02-20 08:32:55.857293794 +0000 UTC m=+6732.628582772" Feb 20 08:33:00 crc kubenswrapper[4492]: I0220 08:33:00.142741 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:33:00 crc kubenswrapper[4492]: I0220 08:33:00.143549 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:33:00 crc kubenswrapper[4492]: I0220 08:33:00.186226 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:33:00 crc kubenswrapper[4492]: I0220 08:33:00.933807 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:33:00 crc kubenswrapper[4492]: I0220 08:33:00.991752 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mp4xc"] Feb 20 08:33:02 crc kubenswrapper[4492]: I0220 08:33:02.907591 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mp4xc" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="registry-server" containerID="cri-o://0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79" gracePeriod=2 Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.401904 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.595667 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2h4x\" (UniqueName: \"kubernetes.io/projected/a82bdc05-86e1-4ae6-a968-30852271a2dc-kube-api-access-j2h4x\") pod \"a82bdc05-86e1-4ae6-a968-30852271a2dc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.595820 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-utilities\") pod \"a82bdc05-86e1-4ae6-a968-30852271a2dc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.596357 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-utilities" (OuterVolumeSpecName: "utilities") pod "a82bdc05-86e1-4ae6-a968-30852271a2dc" (UID: "a82bdc05-86e1-4ae6-a968-30852271a2dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.596606 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-catalog-content\") pod \"a82bdc05-86e1-4ae6-a968-30852271a2dc\" (UID: \"a82bdc05-86e1-4ae6-a968-30852271a2dc\") " Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.601826 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a82bdc05-86e1-4ae6-a968-30852271a2dc-kube-api-access-j2h4x" (OuterVolumeSpecName: "kube-api-access-j2h4x") pod "a82bdc05-86e1-4ae6-a968-30852271a2dc" (UID: "a82bdc05-86e1-4ae6-a968-30852271a2dc"). InnerVolumeSpecName "kube-api-access-j2h4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.603604 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.640448 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a82bdc05-86e1-4ae6-a968-30852271a2dc" (UID: "a82bdc05-86e1-4ae6-a968-30852271a2dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.708656 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2h4x\" (UniqueName: \"kubernetes.io/projected/a82bdc05-86e1-4ae6-a968-30852271a2dc-kube-api-access-j2h4x\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.709178 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a82bdc05-86e1-4ae6-a968-30852271a2dc-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.921539 4492 generic.go:334] "Generic (PLEG): container finished" podID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerID="0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79" exitCode=0 Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.921596 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp4xc" event={"ID":"a82bdc05-86e1-4ae6-a968-30852271a2dc","Type":"ContainerDied","Data":"0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79"} Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.921620 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mp4xc" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.921640 4492 scope.go:117] "RemoveContainer" containerID="0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.921630 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mp4xc" event={"ID":"a82bdc05-86e1-4ae6-a968-30852271a2dc","Type":"ContainerDied","Data":"33526abd88a84d6472950584b2dc8faf2bb21615dae17f2aedcbb7cc991935fe"} Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.956398 4492 scope.go:117] "RemoveContainer" containerID="e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5" Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.959420 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mp4xc"] Feb 20 08:33:03 crc kubenswrapper[4492]: I0220 08:33:03.967648 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mp4xc"] Feb 20 08:33:04 crc kubenswrapper[4492]: I0220 08:33:04.000509 4492 scope.go:117] "RemoveContainer" containerID="603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c" Feb 20 08:33:04 crc kubenswrapper[4492]: I0220 08:33:04.024806 4492 scope.go:117] "RemoveContainer" containerID="0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79" Feb 20 08:33:04 crc kubenswrapper[4492]: E0220 08:33:04.025254 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79\": container with ID starting with 0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79 not found: ID does not exist" containerID="0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79" Feb 20 08:33:04 crc kubenswrapper[4492]: I0220 08:33:04.025295 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79"} err="failed to get container status \"0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79\": rpc error: code = NotFound desc = could not find container \"0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79\": container with ID starting with 0c7d7a78c55e18f3c481077276d969961d431d01dac452c6aa653386f3458c79 not found: ID does not exist" Feb 20 08:33:04 crc kubenswrapper[4492]: I0220 08:33:04.025318 4492 scope.go:117] "RemoveContainer" containerID="e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5" Feb 20 08:33:04 crc kubenswrapper[4492]: E0220 08:33:04.025790 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5\": container with ID starting with e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5 not found: ID does not exist" containerID="e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5" Feb 20 08:33:04 crc kubenswrapper[4492]: I0220 08:33:04.025974 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5"} err="failed to get container status \"e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5\": rpc error: code = NotFound desc = could not find container \"e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5\": container with ID starting with e87fdf5f18bed9387d68b774f89d21469ca42ba784ef776356ea54f3617629b5 not found: ID does not exist" Feb 20 08:33:04 crc kubenswrapper[4492]: I0220 08:33:04.026008 4492 scope.go:117] "RemoveContainer" containerID="603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c" Feb 20 08:33:04 crc kubenswrapper[4492]: E0220 08:33:04.026335 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c\": container with ID starting with 603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c not found: ID does not exist" containerID="603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c" Feb 20 08:33:04 crc kubenswrapper[4492]: I0220 08:33:04.026354 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c"} err="failed to get container status \"603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c\": rpc error: code = NotFound desc = could not find container \"603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c\": container with ID starting with 603779b0701c0ec3db51bb2c2f67ed8f4832ff75daa29e864530f47dc121c72c not found: ID does not exist" Feb 20 08:33:05 crc kubenswrapper[4492]: I0220 08:33:05.566203 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" path="/var/lib/kubelet/pods/a82bdc05-86e1-4ae6-a968-30852271a2dc/volumes" Feb 20 08:33:39 crc kubenswrapper[4492]: I0220 08:33:39.311737 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:33:39 crc kubenswrapper[4492]: I0220 08:33:39.312467 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:34:09 crc kubenswrapper[4492]: I0220 08:34:09.311033 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:34:09 crc kubenswrapper[4492]: I0220 08:34:09.311679 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.311350 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.312084 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.312148 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.312807 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e98f704439beef93eb8425535fe35b539ac237fdf8b7df91e9e87c9064a354a"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.312870 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://1e98f704439beef93eb8425535fe35b539ac237fdf8b7df91e9e87c9064a354a" gracePeriod=600 Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.883604 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="1e98f704439beef93eb8425535fe35b539ac237fdf8b7df91e9e87c9064a354a" exitCode=0 Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.883697 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"1e98f704439beef93eb8425535fe35b539ac237fdf8b7df91e9e87c9064a354a"} Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.883962 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362"} Feb 20 08:34:39 crc kubenswrapper[4492]: I0220 08:34:39.883999 4492 scope.go:117] "RemoveContainer" containerID="4e4d41790f215dd3a07e9c4ee346688835f9d4418e469b3c20a99f53e3142de1" Feb 20 08:36:39 crc kubenswrapper[4492]: I0220 08:36:39.311522 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:36:39 crc kubenswrapper[4492]: I0220 08:36:39.312269 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:37:09 crc kubenswrapper[4492]: I0220 08:37:09.311743 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:37:09 crc kubenswrapper[4492]: I0220 08:37:09.312441 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.790028 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-msh4b"] Feb 20 08:37:36 crc kubenswrapper[4492]: E0220 08:37:36.790867 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="registry-server" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.790881 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="registry-server" Feb 20 08:37:36 crc kubenswrapper[4492]: E0220 08:37:36.790909 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="extract-content" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.790914 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="extract-content" Feb 20 08:37:36 crc kubenswrapper[4492]: E0220 08:37:36.790929 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="extract-utilities" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.790936 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="extract-utilities" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.791090 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82bdc05-86e1-4ae6-a968-30852271a2dc" containerName="registry-server" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.794284 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.815335 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-msh4b"] Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.833813 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmd9l\" (UniqueName: \"kubernetes.io/projected/923371ab-8b58-4654-a954-cc47bf31adb6-kube-api-access-vmd9l\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.833893 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-catalog-content\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.834021 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-utilities\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.935955 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmd9l\" (UniqueName: \"kubernetes.io/projected/923371ab-8b58-4654-a954-cc47bf31adb6-kube-api-access-vmd9l\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.936333 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-catalog-content\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.936489 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-utilities\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.936789 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-catalog-content\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.936929 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-utilities\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:36 crc kubenswrapper[4492]: I0220 08:37:36.955809 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmd9l\" (UniqueName: \"kubernetes.io/projected/923371ab-8b58-4654-a954-cc47bf31adb6-kube-api-access-vmd9l\") pod \"community-operators-msh4b\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:37 crc kubenswrapper[4492]: I0220 08:37:37.110963 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:37 crc kubenswrapper[4492]: I0220 08:37:37.602736 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-msh4b"] Feb 20 08:37:37 crc kubenswrapper[4492]: I0220 08:37:37.641537 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msh4b" event={"ID":"923371ab-8b58-4654-a954-cc47bf31adb6","Type":"ContainerStarted","Data":"64c983f3a7299f810203d0ec929bbba4ab7dc1f2acdbd52f8d95c9e552d793b9"} Feb 20 08:37:38 crc kubenswrapper[4492]: I0220 08:37:38.653133 4492 generic.go:334] "Generic (PLEG): container finished" podID="923371ab-8b58-4654-a954-cc47bf31adb6" containerID="fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb" exitCode=0 Feb 20 08:37:38 crc kubenswrapper[4492]: I0220 08:37:38.653238 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msh4b" event={"ID":"923371ab-8b58-4654-a954-cc47bf31adb6","Type":"ContainerDied","Data":"fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb"} Feb 20 08:37:38 crc kubenswrapper[4492]: I0220 08:37:38.657029 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.311074 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.311503 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.311593 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.312831 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.312919 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" gracePeriod=600 Feb 20 08:37:39 crc kubenswrapper[4492]: E0220 08:37:39.435215 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.663495 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" exitCode=0 Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.663506 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362"} Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.663607 4492 scope.go:117] "RemoveContainer" containerID="1e98f704439beef93eb8425535fe35b539ac237fdf8b7df91e9e87c9064a354a" Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.664367 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:37:39 crc kubenswrapper[4492]: E0220 08:37:39.664745 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:37:39 crc kubenswrapper[4492]: I0220 08:37:39.666005 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msh4b" event={"ID":"923371ab-8b58-4654-a954-cc47bf31adb6","Type":"ContainerStarted","Data":"b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f"} Feb 20 08:37:40 crc kubenswrapper[4492]: I0220 08:37:40.683563 4492 generic.go:334] "Generic (PLEG): container finished" podID="923371ab-8b58-4654-a954-cc47bf31adb6" containerID="b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f" exitCode=0 Feb 20 08:37:40 crc kubenswrapper[4492]: I0220 08:37:40.684424 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msh4b" event={"ID":"923371ab-8b58-4654-a954-cc47bf31adb6","Type":"ContainerDied","Data":"b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f"} Feb 20 08:37:41 crc kubenswrapper[4492]: I0220 08:37:41.716119 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msh4b" event={"ID":"923371ab-8b58-4654-a954-cc47bf31adb6","Type":"ContainerStarted","Data":"aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217"} Feb 20 08:37:41 crc kubenswrapper[4492]: I0220 08:37:41.740784 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-msh4b" podStartSLOduration=3.146795398 podStartE2EDuration="5.740766219s" podCreationTimestamp="2026-02-20 08:37:36 +0000 UTC" firstStartedPulling="2026-02-20 08:37:38.655339755 +0000 UTC m=+7015.426628733" lastFinishedPulling="2026-02-20 08:37:41.249310577 +0000 UTC m=+7018.020599554" observedRunningTime="2026-02-20 08:37:41.738763813 +0000 UTC m=+7018.510052791" watchObservedRunningTime="2026-02-20 08:37:41.740766219 +0000 UTC m=+7018.512055197" Feb 20 08:37:47 crc kubenswrapper[4492]: I0220 08:37:47.111273 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:47 crc kubenswrapper[4492]: I0220 08:37:47.111981 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:47 crc kubenswrapper[4492]: I0220 08:37:47.154614 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:47 crc kubenswrapper[4492]: I0220 08:37:47.826970 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:47 crc kubenswrapper[4492]: I0220 08:37:47.876874 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-msh4b"] Feb 20 08:37:49 crc kubenswrapper[4492]: I0220 08:37:49.791036 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-msh4b" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="registry-server" containerID="cri-o://aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217" gracePeriod=2 Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.336824 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.367329 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-utilities\") pod \"923371ab-8b58-4654-a954-cc47bf31adb6\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.367528 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmd9l\" (UniqueName: \"kubernetes.io/projected/923371ab-8b58-4654-a954-cc47bf31adb6-kube-api-access-vmd9l\") pod \"923371ab-8b58-4654-a954-cc47bf31adb6\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.367625 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-catalog-content\") pod \"923371ab-8b58-4654-a954-cc47bf31adb6\" (UID: \"923371ab-8b58-4654-a954-cc47bf31adb6\") " Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.368254 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-utilities" (OuterVolumeSpecName: "utilities") pod "923371ab-8b58-4654-a954-cc47bf31adb6" (UID: "923371ab-8b58-4654-a954-cc47bf31adb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.378120 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/923371ab-8b58-4654-a954-cc47bf31adb6-kube-api-access-vmd9l" (OuterVolumeSpecName: "kube-api-access-vmd9l") pod "923371ab-8b58-4654-a954-cc47bf31adb6" (UID: "923371ab-8b58-4654-a954-cc47bf31adb6"). InnerVolumeSpecName "kube-api-access-vmd9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.425968 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "923371ab-8b58-4654-a954-cc47bf31adb6" (UID: "923371ab-8b58-4654-a954-cc47bf31adb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.469677 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.469711 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/923371ab-8b58-4654-a954-cc47bf31adb6-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.469724 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmd9l\" (UniqueName: \"kubernetes.io/projected/923371ab-8b58-4654-a954-cc47bf31adb6-kube-api-access-vmd9l\") on node \"crc\" DevicePath \"\"" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.557410 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:37:50 crc kubenswrapper[4492]: E0220 08:37:50.557758 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.801706 4492 generic.go:334] "Generic (PLEG): container finished" podID="923371ab-8b58-4654-a954-cc47bf31adb6" containerID="aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217" exitCode=0 Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.801770 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msh4b" event={"ID":"923371ab-8b58-4654-a954-cc47bf31adb6","Type":"ContainerDied","Data":"aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217"} Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.801803 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-msh4b" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.801832 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-msh4b" event={"ID":"923371ab-8b58-4654-a954-cc47bf31adb6","Type":"ContainerDied","Data":"64c983f3a7299f810203d0ec929bbba4ab7dc1f2acdbd52f8d95c9e552d793b9"} Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.801862 4492 scope.go:117] "RemoveContainer" containerID="aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.826393 4492 scope.go:117] "RemoveContainer" containerID="b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.854505 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-msh4b"] Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.868460 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-msh4b"] Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.873625 4492 scope.go:117] "RemoveContainer" containerID="fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.917942 4492 scope.go:117] "RemoveContainer" containerID="aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217" Feb 20 08:37:50 crc kubenswrapper[4492]: E0220 08:37:50.918493 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217\": container with ID starting with aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217 not found: ID does not exist" containerID="aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.918541 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217"} err="failed to get container status \"aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217\": rpc error: code = NotFound desc = could not find container \"aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217\": container with ID starting with aa6784f350bd4dd71e7687670d411203e8c516d498d051bbb492a4baf20e5217 not found: ID does not exist" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.918576 4492 scope.go:117] "RemoveContainer" containerID="b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f" Feb 20 08:37:50 crc kubenswrapper[4492]: E0220 08:37:50.918924 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f\": container with ID starting with b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f not found: ID does not exist" containerID="b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.918963 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f"} err="failed to get container status \"b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f\": rpc error: code = NotFound desc = could not find container \"b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f\": container with ID starting with b3f5969c8dbeef372cb467a0bb659dd6ae46fab2af8280c826b07d2a06c4ab4f not found: ID does not exist" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.918992 4492 scope.go:117] "RemoveContainer" containerID="fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb" Feb 20 08:37:50 crc kubenswrapper[4492]: E0220 08:37:50.919291 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb\": container with ID starting with fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb not found: ID does not exist" containerID="fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb" Feb 20 08:37:50 crc kubenswrapper[4492]: I0220 08:37:50.919326 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb"} err="failed to get container status \"fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb\": rpc error: code = NotFound desc = could not find container \"fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb\": container with ID starting with fe413d61a4be67e2c62ea832f5729b686a14157becd1cf39cfdd036a9ef613eb not found: ID does not exist" Feb 20 08:37:51 crc kubenswrapper[4492]: I0220 08:37:51.576122 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" path="/var/lib/kubelet/pods/923371ab-8b58-4654-a954-cc47bf31adb6/volumes" Feb 20 08:38:05 crc kubenswrapper[4492]: I0220 08:38:05.557744 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:38:05 crc kubenswrapper[4492]: E0220 08:38:05.558752 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:38:19 crc kubenswrapper[4492]: I0220 08:38:19.558147 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:38:19 crc kubenswrapper[4492]: E0220 08:38:19.559298 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:38:31 crc kubenswrapper[4492]: I0220 08:38:31.556979 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:38:31 crc kubenswrapper[4492]: E0220 08:38:31.557926 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:38:44 crc kubenswrapper[4492]: I0220 08:38:44.557713 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:38:44 crc kubenswrapper[4492]: E0220 08:38:44.558662 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:38:59 crc kubenswrapper[4492]: I0220 08:38:59.560043 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:38:59 crc kubenswrapper[4492]: E0220 08:38:59.560984 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:39:12 crc kubenswrapper[4492]: I0220 08:39:12.557568 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:39:12 crc kubenswrapper[4492]: E0220 08:39:12.558597 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:39:26 crc kubenswrapper[4492]: I0220 08:39:26.556888 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:39:26 crc kubenswrapper[4492]: E0220 08:39:26.557712 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:39:39 crc kubenswrapper[4492]: I0220 08:39:39.559313 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:39:39 crc kubenswrapper[4492]: E0220 08:39:39.560421 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:39:52 crc kubenswrapper[4492]: I0220 08:39:52.557962 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:39:52 crc kubenswrapper[4492]: E0220 08:39:52.558937 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:40:04 crc kubenswrapper[4492]: I0220 08:40:04.557814 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:40:04 crc kubenswrapper[4492]: E0220 08:40:04.558786 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:40:18 crc kubenswrapper[4492]: I0220 08:40:18.557734 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:40:18 crc kubenswrapper[4492]: E0220 08:40:18.558820 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:40:32 crc kubenswrapper[4492]: I0220 08:40:32.558568 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:40:32 crc kubenswrapper[4492]: E0220 08:40:32.559791 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:40:45 crc kubenswrapper[4492]: I0220 08:40:45.557732 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:40:45 crc kubenswrapper[4492]: E0220 08:40:45.559650 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:40:58 crc kubenswrapper[4492]: I0220 08:40:58.557465 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:40:58 crc kubenswrapper[4492]: E0220 08:40:58.558499 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:41:09 crc kubenswrapper[4492]: I0220 08:41:09.557053 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:41:09 crc kubenswrapper[4492]: E0220 08:41:09.557937 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:41:23 crc kubenswrapper[4492]: I0220 08:41:23.563671 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:41:23 crc kubenswrapper[4492]: E0220 08:41:23.564667 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:41:34 crc kubenswrapper[4492]: I0220 08:41:34.557333 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:41:34 crc kubenswrapper[4492]: E0220 08:41:34.558390 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:41:49 crc kubenswrapper[4492]: I0220 08:41:49.558409 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:41:49 crc kubenswrapper[4492]: E0220 08:41:49.559281 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:42:02 crc kubenswrapper[4492]: I0220 08:42:02.557075 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:42:02 crc kubenswrapper[4492]: E0220 08:42:02.558857 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:42:17 crc kubenswrapper[4492]: I0220 08:42:17.557795 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:42:17 crc kubenswrapper[4492]: E0220 08:42:17.558753 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.340683 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8kssb"] Feb 20 08:42:25 crc kubenswrapper[4492]: E0220 08:42:25.342343 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="extract-utilities" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.342408 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="extract-utilities" Feb 20 08:42:25 crc kubenswrapper[4492]: E0220 08:42:25.342460 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="extract-content" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.342530 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="extract-content" Feb 20 08:42:25 crc kubenswrapper[4492]: E0220 08:42:25.342603 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="registry-server" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.342699 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="registry-server" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.342915 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="923371ab-8b58-4654-a954-cc47bf31adb6" containerName="registry-server" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.345504 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.364236 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kssb"] Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.390423 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-catalog-content\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.390548 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-utilities\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.390629 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7x5l\" (UniqueName: \"kubernetes.io/projected/180e062c-e589-4c0c-9791-1f844d10e951-kube-api-access-d7x5l\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.492617 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-catalog-content\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.492663 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-utilities\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.492690 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7x5l\" (UniqueName: \"kubernetes.io/projected/180e062c-e589-4c0c-9791-1f844d10e951-kube-api-access-d7x5l\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.493164 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-utilities\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.493387 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-catalog-content\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.512528 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7x5l\" (UniqueName: \"kubernetes.io/projected/180e062c-e589-4c0c-9791-1f844d10e951-kube-api-access-d7x5l\") pod \"redhat-marketplace-8kssb\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:25 crc kubenswrapper[4492]: I0220 08:42:25.661127 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:26 crc kubenswrapper[4492]: I0220 08:42:26.330887 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kssb"] Feb 20 08:42:27 crc kubenswrapper[4492]: I0220 08:42:27.288400 4492 generic.go:334] "Generic (PLEG): container finished" podID="180e062c-e589-4c0c-9791-1f844d10e951" containerID="bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a" exitCode=0 Feb 20 08:42:27 crc kubenswrapper[4492]: I0220 08:42:27.288605 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kssb" event={"ID":"180e062c-e589-4c0c-9791-1f844d10e951","Type":"ContainerDied","Data":"bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a"} Feb 20 08:42:27 crc kubenswrapper[4492]: I0220 08:42:27.288844 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kssb" event={"ID":"180e062c-e589-4c0c-9791-1f844d10e951","Type":"ContainerStarted","Data":"2099a75d807f74c80aaa0a5e04b6de8e06219ba2f109ea26e17a851a01d7ac42"} Feb 20 08:42:28 crc kubenswrapper[4492]: I0220 08:42:28.302120 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kssb" event={"ID":"180e062c-e589-4c0c-9791-1f844d10e951","Type":"ContainerStarted","Data":"7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575"} Feb 20 08:42:28 crc kubenswrapper[4492]: E0220 08:42:28.893536 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod180e062c_e589_4c0c_9791_1f844d10e951.slice/crio-7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod180e062c_e589_4c0c_9791_1f844d10e951.slice/crio-conmon-7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575.scope\": RecentStats: unable to find data in memory cache]" Feb 20 08:42:29 crc kubenswrapper[4492]: I0220 08:42:29.314550 4492 generic.go:334] "Generic (PLEG): container finished" podID="180e062c-e589-4c0c-9791-1f844d10e951" containerID="7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575" exitCode=0 Feb 20 08:42:29 crc kubenswrapper[4492]: I0220 08:42:29.314635 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kssb" event={"ID":"180e062c-e589-4c0c-9791-1f844d10e951","Type":"ContainerDied","Data":"7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575"} Feb 20 08:42:30 crc kubenswrapper[4492]: I0220 08:42:30.329608 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kssb" event={"ID":"180e062c-e589-4c0c-9791-1f844d10e951","Type":"ContainerStarted","Data":"bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37"} Feb 20 08:42:30 crc kubenswrapper[4492]: I0220 08:42:30.346211 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8kssb" podStartSLOduration=2.856309147 podStartE2EDuration="5.346174756s" podCreationTimestamp="2026-02-20 08:42:25 +0000 UTC" firstStartedPulling="2026-02-20 08:42:27.291292547 +0000 UTC m=+7304.062581525" lastFinishedPulling="2026-02-20 08:42:29.781158156 +0000 UTC m=+7306.552447134" observedRunningTime="2026-02-20 08:42:30.345990849 +0000 UTC m=+7307.117279827" watchObservedRunningTime="2026-02-20 08:42:30.346174756 +0000 UTC m=+7307.117463734" Feb 20 08:42:32 crc kubenswrapper[4492]: I0220 08:42:32.557394 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:42:32 crc kubenswrapper[4492]: E0220 08:42:32.558047 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:42:35 crc kubenswrapper[4492]: I0220 08:42:35.661229 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:35 crc kubenswrapper[4492]: I0220 08:42:35.661663 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:35 crc kubenswrapper[4492]: I0220 08:42:35.711186 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:36 crc kubenswrapper[4492]: I0220 08:42:36.431424 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:36 crc kubenswrapper[4492]: I0220 08:42:36.475348 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kssb"] Feb 20 08:42:38 crc kubenswrapper[4492]: I0220 08:42:38.402653 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8kssb" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="registry-server" containerID="cri-o://bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37" gracePeriod=2 Feb 20 08:42:38 crc kubenswrapper[4492]: I0220 08:42:38.964549 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.006681 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7x5l\" (UniqueName: \"kubernetes.io/projected/180e062c-e589-4c0c-9791-1f844d10e951-kube-api-access-d7x5l\") pod \"180e062c-e589-4c0c-9791-1f844d10e951\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.007081 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-catalog-content\") pod \"180e062c-e589-4c0c-9791-1f844d10e951\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.007230 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-utilities\") pod \"180e062c-e589-4c0c-9791-1f844d10e951\" (UID: \"180e062c-e589-4c0c-9791-1f844d10e951\") " Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.007867 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-utilities" (OuterVolumeSpecName: "utilities") pod "180e062c-e589-4c0c-9791-1f844d10e951" (UID: "180e062c-e589-4c0c-9791-1f844d10e951"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.008572 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.015563 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180e062c-e589-4c0c-9791-1f844d10e951-kube-api-access-d7x5l" (OuterVolumeSpecName: "kube-api-access-d7x5l") pod "180e062c-e589-4c0c-9791-1f844d10e951" (UID: "180e062c-e589-4c0c-9791-1f844d10e951"). InnerVolumeSpecName "kube-api-access-d7x5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.032515 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "180e062c-e589-4c0c-9791-1f844d10e951" (UID: "180e062c-e589-4c0c-9791-1f844d10e951"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.109947 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7x5l\" (UniqueName: \"kubernetes.io/projected/180e062c-e589-4c0c-9791-1f844d10e951-kube-api-access-d7x5l\") on node \"crc\" DevicePath \"\"" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.109986 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/180e062c-e589-4c0c-9791-1f844d10e951-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.416363 4492 generic.go:334] "Generic (PLEG): container finished" podID="180e062c-e589-4c0c-9791-1f844d10e951" containerID="bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37" exitCode=0 Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.416449 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kssb" event={"ID":"180e062c-e589-4c0c-9791-1f844d10e951","Type":"ContainerDied","Data":"bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37"} Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.416765 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8kssb" event={"ID":"180e062c-e589-4c0c-9791-1f844d10e951","Type":"ContainerDied","Data":"2099a75d807f74c80aaa0a5e04b6de8e06219ba2f109ea26e17a851a01d7ac42"} Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.416467 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8kssb" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.416801 4492 scope.go:117] "RemoveContainer" containerID="bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.463719 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kssb"] Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.468690 4492 scope.go:117] "RemoveContainer" containerID="7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.474003 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8kssb"] Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.490756 4492 scope.go:117] "RemoveContainer" containerID="bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.533879 4492 scope.go:117] "RemoveContainer" containerID="bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37" Feb 20 08:42:39 crc kubenswrapper[4492]: E0220 08:42:39.536503 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37\": container with ID starting with bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37 not found: ID does not exist" containerID="bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.536547 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37"} err="failed to get container status \"bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37\": rpc error: code = NotFound desc = could not find container \"bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37\": container with ID starting with bb4ff325925133214c89d8c5f80dc24edd48efc1ba022797240a22b446064a37 not found: ID does not exist" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.536576 4492 scope.go:117] "RemoveContainer" containerID="7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575" Feb 20 08:42:39 crc kubenswrapper[4492]: E0220 08:42:39.537341 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575\": container with ID starting with 7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575 not found: ID does not exist" containerID="7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.537393 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575"} err="failed to get container status \"7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575\": rpc error: code = NotFound desc = could not find container \"7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575\": container with ID starting with 7c1732d9c893b9e58782857baeae60b3401e40d862423b3ab94e32634e87c575 not found: ID does not exist" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.537426 4492 scope.go:117] "RemoveContainer" containerID="bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a" Feb 20 08:42:39 crc kubenswrapper[4492]: E0220 08:42:39.537750 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a\": container with ID starting with bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a not found: ID does not exist" containerID="bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.537802 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a"} err="failed to get container status \"bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a\": rpc error: code = NotFound desc = could not find container \"bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a\": container with ID starting with bd99f27af8a4ec0497eec2d2301cfb7a13473aad2262310d14637a6d15cac63a not found: ID does not exist" Feb 20 08:42:39 crc kubenswrapper[4492]: I0220 08:42:39.571156 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="180e062c-e589-4c0c-9791-1f844d10e951" path="/var/lib/kubelet/pods/180e062c-e589-4c0c-9791-1f844d10e951/volumes" Feb 20 08:42:47 crc kubenswrapper[4492]: I0220 08:42:47.556868 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:42:48 crc kubenswrapper[4492]: I0220 08:42:48.499272 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"ed761079a9284278df7d30a279bc0b5c625e1e30cacd794793b0f9fbe570624d"} Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.113389 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-784sk"] Feb 20 08:42:57 crc kubenswrapper[4492]: E0220 08:42:57.115721 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="extract-utilities" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.115846 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="extract-utilities" Feb 20 08:42:57 crc kubenswrapper[4492]: E0220 08:42:57.115919 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="extract-content" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.115973 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="extract-content" Feb 20 08:42:57 crc kubenswrapper[4492]: E0220 08:42:57.116043 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="registry-server" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.116090 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="registry-server" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.116394 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="180e062c-e589-4c0c-9791-1f844d10e951" containerName="registry-server" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.118046 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.122659 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-784sk"] Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.207547 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-catalog-content\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.207674 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twwfs\" (UniqueName: \"kubernetes.io/projected/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-kube-api-access-twwfs\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.207743 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-utilities\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.310969 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twwfs\" (UniqueName: \"kubernetes.io/projected/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-kube-api-access-twwfs\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.311070 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-utilities\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.311130 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-catalog-content\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.311773 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-utilities\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.311850 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-catalog-content\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.332346 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twwfs\" (UniqueName: \"kubernetes.io/projected/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-kube-api-access-twwfs\") pod \"redhat-operators-784sk\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.435213 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:42:57 crc kubenswrapper[4492]: I0220 08:42:57.899665 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-784sk"] Feb 20 08:42:58 crc kubenswrapper[4492]: I0220 08:42:58.595987 4492 generic.go:334] "Generic (PLEG): container finished" podID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerID="2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023" exitCode=0 Feb 20 08:42:58 crc kubenswrapper[4492]: I0220 08:42:58.596093 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-784sk" event={"ID":"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a","Type":"ContainerDied","Data":"2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023"} Feb 20 08:42:58 crc kubenswrapper[4492]: I0220 08:42:58.596358 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-784sk" event={"ID":"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a","Type":"ContainerStarted","Data":"e38f5d181040ce6868e58a2cdc08814e0e9487a9ebf7929734e4e36b8a680493"} Feb 20 08:42:58 crc kubenswrapper[4492]: I0220 08:42:58.597881 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:42:59 crc kubenswrapper[4492]: I0220 08:42:59.642720 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-784sk" event={"ID":"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a","Type":"ContainerStarted","Data":"8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f"} Feb 20 08:43:02 crc kubenswrapper[4492]: I0220 08:43:02.671756 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-784sk" event={"ID":"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a","Type":"ContainerDied","Data":"8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f"} Feb 20 08:43:02 crc kubenswrapper[4492]: I0220 08:43:02.672520 4492 generic.go:334] "Generic (PLEG): container finished" podID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerID="8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f" exitCode=0 Feb 20 08:43:03 crc kubenswrapper[4492]: I0220 08:43:03.684576 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-784sk" event={"ID":"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a","Type":"ContainerStarted","Data":"d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609"} Feb 20 08:43:03 crc kubenswrapper[4492]: I0220 08:43:03.710349 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-784sk" podStartSLOduration=2.149757791 podStartE2EDuration="6.710321013s" podCreationTimestamp="2026-02-20 08:42:57 +0000 UTC" firstStartedPulling="2026-02-20 08:42:58.597628562 +0000 UTC m=+7335.368917539" lastFinishedPulling="2026-02-20 08:43:03.158191783 +0000 UTC m=+7339.929480761" observedRunningTime="2026-02-20 08:43:03.700814597 +0000 UTC m=+7340.472103576" watchObservedRunningTime="2026-02-20 08:43:03.710321013 +0000 UTC m=+7340.481609991" Feb 20 08:43:07 crc kubenswrapper[4492]: I0220 08:43:07.435678 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:43:07 crc kubenswrapper[4492]: I0220 08:43:07.436417 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:43:08 crc kubenswrapper[4492]: I0220 08:43:08.477190 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-784sk" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="registry-server" probeResult="failure" output=< Feb 20 08:43:08 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:43:08 crc kubenswrapper[4492]: > Feb 20 08:43:17 crc kubenswrapper[4492]: I0220 08:43:17.478839 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:43:17 crc kubenswrapper[4492]: I0220 08:43:17.543561 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:43:17 crc kubenswrapper[4492]: I0220 08:43:17.729019 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-784sk"] Feb 20 08:43:18 crc kubenswrapper[4492]: I0220 08:43:18.821454 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-784sk" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="registry-server" containerID="cri-o://d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609" gracePeriod=2 Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.419976 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.501180 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-utilities\") pod \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.501243 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-catalog-content\") pod \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.501424 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twwfs\" (UniqueName: \"kubernetes.io/projected/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-kube-api-access-twwfs\") pod \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\" (UID: \"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a\") " Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.502456 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-utilities" (OuterVolumeSpecName: "utilities") pod "5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" (UID: "5b08dfa5-f319-4efb-a222-03ae6c8ccd7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.510563 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-kube-api-access-twwfs" (OuterVolumeSpecName: "kube-api-access-twwfs") pod "5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" (UID: "5b08dfa5-f319-4efb-a222-03ae6c8ccd7a"). InnerVolumeSpecName "kube-api-access-twwfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.604042 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.604091 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twwfs\" (UniqueName: \"kubernetes.io/projected/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-kube-api-access-twwfs\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.610966 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" (UID: "5b08dfa5-f319-4efb-a222-03ae6c8ccd7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.704891 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.832270 4492 generic.go:334] "Generic (PLEG): container finished" podID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerID="d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609" exitCode=0 Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.832323 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-784sk" event={"ID":"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a","Type":"ContainerDied","Data":"d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609"} Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.832333 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-784sk" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.832359 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-784sk" event={"ID":"5b08dfa5-f319-4efb-a222-03ae6c8ccd7a","Type":"ContainerDied","Data":"e38f5d181040ce6868e58a2cdc08814e0e9487a9ebf7929734e4e36b8a680493"} Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.832403 4492 scope.go:117] "RemoveContainer" containerID="d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.863185 4492 scope.go:117] "RemoveContainer" containerID="8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.874190 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-784sk"] Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.881132 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-784sk"] Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.885652 4492 scope.go:117] "RemoveContainer" containerID="2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.926385 4492 scope.go:117] "RemoveContainer" containerID="d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609" Feb 20 08:43:19 crc kubenswrapper[4492]: E0220 08:43:19.927556 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609\": container with ID starting with d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609 not found: ID does not exist" containerID="d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.928226 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609"} err="failed to get container status \"d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609\": rpc error: code = NotFound desc = could not find container \"d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609\": container with ID starting with d002222e145958d9998256b7323378f4f0d366e8d1dd9acd15236614fc649609 not found: ID does not exist" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.928262 4492 scope.go:117] "RemoveContainer" containerID="8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f" Feb 20 08:43:19 crc kubenswrapper[4492]: E0220 08:43:19.928546 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f\": container with ID starting with 8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f not found: ID does not exist" containerID="8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.928627 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f"} err="failed to get container status \"8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f\": rpc error: code = NotFound desc = could not find container \"8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f\": container with ID starting with 8b0f762aadc59637b11a8827f076073817a772935f5cb80c00fc81521516036f not found: ID does not exist" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.928694 4492 scope.go:117] "RemoveContainer" containerID="2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023" Feb 20 08:43:19 crc kubenswrapper[4492]: E0220 08:43:19.929402 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023\": container with ID starting with 2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023 not found: ID does not exist" containerID="2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023" Feb 20 08:43:19 crc kubenswrapper[4492]: I0220 08:43:19.929455 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023"} err="failed to get container status \"2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023\": rpc error: code = NotFound desc = could not find container \"2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023\": container with ID starting with 2875f31a67d4f7efe57bcd4916bfeee3cfaa6edf8d567a747aebb03468c3c023 not found: ID does not exist" Feb 20 08:43:21 crc kubenswrapper[4492]: I0220 08:43:21.566992 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" path="/var/lib/kubelet/pods/5b08dfa5-f319-4efb-a222-03ae6c8ccd7a/volumes" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.751447 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6lc22"] Feb 20 08:43:56 crc kubenswrapper[4492]: E0220 08:43:56.752604 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="extract-content" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.752619 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="extract-content" Feb 20 08:43:56 crc kubenswrapper[4492]: E0220 08:43:56.752646 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="registry-server" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.752653 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="registry-server" Feb 20 08:43:56 crc kubenswrapper[4492]: E0220 08:43:56.752672 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="extract-utilities" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.752678 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="extract-utilities" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.752861 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b08dfa5-f319-4efb-a222-03ae6c8ccd7a" containerName="registry-server" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.754162 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.778519 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6lc22"] Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.829449 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-catalog-content\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.829916 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wprvc\" (UniqueName: \"kubernetes.io/projected/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-kube-api-access-wprvc\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.830135 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-utilities\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.931597 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-catalog-content\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.931675 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wprvc\" (UniqueName: \"kubernetes.io/projected/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-kube-api-access-wprvc\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.931779 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-utilities\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.932151 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-catalog-content\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.932245 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-utilities\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:56 crc kubenswrapper[4492]: I0220 08:43:56.958218 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wprvc\" (UniqueName: \"kubernetes.io/projected/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-kube-api-access-wprvc\") pod \"certified-operators-6lc22\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:57 crc kubenswrapper[4492]: I0220 08:43:57.071957 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:43:57 crc kubenswrapper[4492]: I0220 08:43:57.515618 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6lc22"] Feb 20 08:43:58 crc kubenswrapper[4492]: I0220 08:43:58.139733 4492 generic.go:334] "Generic (PLEG): container finished" podID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerID="2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88" exitCode=0 Feb 20 08:43:58 crc kubenswrapper[4492]: I0220 08:43:58.139946 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lc22" event={"ID":"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5","Type":"ContainerDied","Data":"2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88"} Feb 20 08:43:58 crc kubenswrapper[4492]: I0220 08:43:58.140064 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lc22" event={"ID":"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5","Type":"ContainerStarted","Data":"dfa6d2c22aec8fab4ed2eaae8505830eb0e19732111d48d6250f742b7d841e2b"} Feb 20 08:43:59 crc kubenswrapper[4492]: I0220 08:43:59.161941 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lc22" event={"ID":"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5","Type":"ContainerStarted","Data":"691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96"} Feb 20 08:44:01 crc kubenswrapper[4492]: I0220 08:44:01.177406 4492 generic.go:334] "Generic (PLEG): container finished" podID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerID="691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96" exitCode=0 Feb 20 08:44:01 crc kubenswrapper[4492]: I0220 08:44:01.177502 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lc22" event={"ID":"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5","Type":"ContainerDied","Data":"691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96"} Feb 20 08:44:02 crc kubenswrapper[4492]: I0220 08:44:02.187313 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lc22" event={"ID":"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5","Type":"ContainerStarted","Data":"14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a"} Feb 20 08:44:02 crc kubenswrapper[4492]: I0220 08:44:02.208388 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6lc22" podStartSLOduration=2.645983883 podStartE2EDuration="6.208370266s" podCreationTimestamp="2026-02-20 08:43:56 +0000 UTC" firstStartedPulling="2026-02-20 08:43:58.141639838 +0000 UTC m=+7394.912928816" lastFinishedPulling="2026-02-20 08:44:01.704026222 +0000 UTC m=+7398.475315199" observedRunningTime="2026-02-20 08:44:02.202314879 +0000 UTC m=+7398.973603856" watchObservedRunningTime="2026-02-20 08:44:02.208370266 +0000 UTC m=+7398.979659244" Feb 20 08:44:07 crc kubenswrapper[4492]: I0220 08:44:07.072800 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:44:07 crc kubenswrapper[4492]: I0220 08:44:07.073504 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:44:07 crc kubenswrapper[4492]: I0220 08:44:07.112903 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:44:07 crc kubenswrapper[4492]: I0220 08:44:07.260517 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:44:07 crc kubenswrapper[4492]: I0220 08:44:07.346759 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6lc22"] Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.237960 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6lc22" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="registry-server" containerID="cri-o://14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a" gracePeriod=2 Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.663514 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.756228 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-catalog-content\") pod \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.756378 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-utilities\") pod \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.756406 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wprvc\" (UniqueName: \"kubernetes.io/projected/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-kube-api-access-wprvc\") pod \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\" (UID: \"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5\") " Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.759159 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-utilities" (OuterVolumeSpecName: "utilities") pod "a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" (UID: "a8eadfc6-9789-477e-b671-a3d7f5f0cfd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.765212 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-kube-api-access-wprvc" (OuterVolumeSpecName: "kube-api-access-wprvc") pod "a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" (UID: "a8eadfc6-9789-477e-b671-a3d7f5f0cfd5"). InnerVolumeSpecName "kube-api-access-wprvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.804205 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" (UID: "a8eadfc6-9789-477e-b671-a3d7f5f0cfd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.859937 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.860083 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:09 crc kubenswrapper[4492]: I0220 08:44:09.860140 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wprvc\" (UniqueName: \"kubernetes.io/projected/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5-kube-api-access-wprvc\") on node \"crc\" DevicePath \"\"" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.247042 4492 generic.go:334] "Generic (PLEG): container finished" podID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerID="14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a" exitCode=0 Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.247137 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6lc22" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.247136 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lc22" event={"ID":"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5","Type":"ContainerDied","Data":"14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a"} Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.247422 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6lc22" event={"ID":"a8eadfc6-9789-477e-b671-a3d7f5f0cfd5","Type":"ContainerDied","Data":"dfa6d2c22aec8fab4ed2eaae8505830eb0e19732111d48d6250f742b7d841e2b"} Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.247444 4492 scope.go:117] "RemoveContainer" containerID="14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.267148 4492 scope.go:117] "RemoveContainer" containerID="691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.281736 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6lc22"] Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.295994 4492 scope.go:117] "RemoveContainer" containerID="2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.299524 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6lc22"] Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.328593 4492 scope.go:117] "RemoveContainer" containerID="14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a" Feb 20 08:44:10 crc kubenswrapper[4492]: E0220 08:44:10.329035 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a\": container with ID starting with 14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a not found: ID does not exist" containerID="14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.329082 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a"} err="failed to get container status \"14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a\": rpc error: code = NotFound desc = could not find container \"14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a\": container with ID starting with 14c16511da744b6076b5eb0b7e389d5ab593333e3b122449741654a129b1cc5a not found: ID does not exist" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.329111 4492 scope.go:117] "RemoveContainer" containerID="691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96" Feb 20 08:44:10 crc kubenswrapper[4492]: E0220 08:44:10.329538 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96\": container with ID starting with 691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96 not found: ID does not exist" containerID="691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.329581 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96"} err="failed to get container status \"691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96\": rpc error: code = NotFound desc = could not find container \"691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96\": container with ID starting with 691348dbb1ce986f3794e903e308338907378d16e93d731f38c608936574cb96 not found: ID does not exist" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.329617 4492 scope.go:117] "RemoveContainer" containerID="2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88" Feb 20 08:44:10 crc kubenswrapper[4492]: E0220 08:44:10.330043 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88\": container with ID starting with 2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88 not found: ID does not exist" containerID="2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88" Feb 20 08:44:10 crc kubenswrapper[4492]: I0220 08:44:10.330066 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88"} err="failed to get container status \"2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88\": rpc error: code = NotFound desc = could not find container \"2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88\": container with ID starting with 2e7f0a92e468b2a7538ed7e98c9427523522053f17fd7557d70eca6a6e94ef88 not found: ID does not exist" Feb 20 08:44:11 crc kubenswrapper[4492]: I0220 08:44:11.566301 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" path="/var/lib/kubelet/pods/a8eadfc6-9789-477e-b671-a3d7f5f0cfd5/volumes" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.219286 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr"] Feb 20 08:45:00 crc kubenswrapper[4492]: E0220 08:45:00.220503 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="extract-content" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.220521 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="extract-content" Feb 20 08:45:00 crc kubenswrapper[4492]: E0220 08:45:00.220539 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="registry-server" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.220545 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="registry-server" Feb 20 08:45:00 crc kubenswrapper[4492]: E0220 08:45:00.220567 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="extract-utilities" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.220573 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="extract-utilities" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.220820 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8eadfc6-9789-477e-b671-a3d7f5f0cfd5" containerName="registry-server" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.221500 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.232279 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr"] Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.306811 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.306818 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.324049 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvj25\" (UniqueName: \"kubernetes.io/projected/d036832f-1f0c-4f37-be93-5173bc06e81f-kube-api-access-xvj25\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.324172 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d036832f-1f0c-4f37-be93-5173bc06e81f-secret-volume\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.324262 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d036832f-1f0c-4f37-be93-5173bc06e81f-config-volume\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.426587 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d036832f-1f0c-4f37-be93-5173bc06e81f-secret-volume\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.426763 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d036832f-1f0c-4f37-be93-5173bc06e81f-config-volume\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.426897 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvj25\" (UniqueName: \"kubernetes.io/projected/d036832f-1f0c-4f37-be93-5173bc06e81f-kube-api-access-xvj25\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.427758 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d036832f-1f0c-4f37-be93-5173bc06e81f-config-volume\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.434362 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d036832f-1f0c-4f37-be93-5173bc06e81f-secret-volume\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.443244 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvj25\" (UniqueName: \"kubernetes.io/projected/d036832f-1f0c-4f37-be93-5173bc06e81f-kube-api-access-xvj25\") pod \"collect-profiles-29526285-f58xr\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.553097 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:00 crc kubenswrapper[4492]: I0220 08:45:00.980997 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr"] Feb 20 08:45:01 crc kubenswrapper[4492]: I0220 08:45:01.634403 4492 generic.go:334] "Generic (PLEG): container finished" podID="d036832f-1f0c-4f37-be93-5173bc06e81f" containerID="37e4936de6f5566e8270e0e4b05fca9f2e01d7aaff12d46661ba239ac660d20c" exitCode=0 Feb 20 08:45:01 crc kubenswrapper[4492]: I0220 08:45:01.634553 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" event={"ID":"d036832f-1f0c-4f37-be93-5173bc06e81f","Type":"ContainerDied","Data":"37e4936de6f5566e8270e0e4b05fca9f2e01d7aaff12d46661ba239ac660d20c"} Feb 20 08:45:01 crc kubenswrapper[4492]: I0220 08:45:01.634850 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" event={"ID":"d036832f-1f0c-4f37-be93-5173bc06e81f","Type":"ContainerStarted","Data":"55fbbcaf86e36dea41e4d8f61797b4e9bd0eac5ff107d710022d57ae18dcd924"} Feb 20 08:45:02 crc kubenswrapper[4492]: I0220 08:45:02.942040 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.006823 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d036832f-1f0c-4f37-be93-5173bc06e81f-config-volume\") pod \"d036832f-1f0c-4f37-be93-5173bc06e81f\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.007115 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvj25\" (UniqueName: \"kubernetes.io/projected/d036832f-1f0c-4f37-be93-5173bc06e81f-kube-api-access-xvj25\") pod \"d036832f-1f0c-4f37-be93-5173bc06e81f\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.007239 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d036832f-1f0c-4f37-be93-5173bc06e81f-secret-volume\") pod \"d036832f-1f0c-4f37-be93-5173bc06e81f\" (UID: \"d036832f-1f0c-4f37-be93-5173bc06e81f\") " Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.009224 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d036832f-1f0c-4f37-be93-5173bc06e81f-config-volume" (OuterVolumeSpecName: "config-volume") pod "d036832f-1f0c-4f37-be93-5173bc06e81f" (UID: "d036832f-1f0c-4f37-be93-5173bc06e81f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.018715 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d036832f-1f0c-4f37-be93-5173bc06e81f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d036832f-1f0c-4f37-be93-5173bc06e81f" (UID: "d036832f-1f0c-4f37-be93-5173bc06e81f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.044868 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d036832f-1f0c-4f37-be93-5173bc06e81f-kube-api-access-xvj25" (OuterVolumeSpecName: "kube-api-access-xvj25") pod "d036832f-1f0c-4f37-be93-5173bc06e81f" (UID: "d036832f-1f0c-4f37-be93-5173bc06e81f"). InnerVolumeSpecName "kube-api-access-xvj25". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.110770 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvj25\" (UniqueName: \"kubernetes.io/projected/d036832f-1f0c-4f37-be93-5173bc06e81f-kube-api-access-xvj25\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.110823 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d036832f-1f0c-4f37-be93-5173bc06e81f-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.110835 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d036832f-1f0c-4f37-be93-5173bc06e81f-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.654833 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" event={"ID":"d036832f-1f0c-4f37-be93-5173bc06e81f","Type":"ContainerDied","Data":"55fbbcaf86e36dea41e4d8f61797b4e9bd0eac5ff107d710022d57ae18dcd924"} Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.655186 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55fbbcaf86e36dea41e4d8f61797b4e9bd0eac5ff107d710022d57ae18dcd924" Feb 20 08:45:03 crc kubenswrapper[4492]: I0220 08:45:03.654893 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr" Feb 20 08:45:04 crc kubenswrapper[4492]: I0220 08:45:04.031802 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh"] Feb 20 08:45:04 crc kubenswrapper[4492]: I0220 08:45:04.038400 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526240-rjlvh"] Feb 20 08:45:05 crc kubenswrapper[4492]: I0220 08:45:05.576321 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b2c7017-b764-452f-9783-c8ab0b90fbdc" path="/var/lib/kubelet/pods/3b2c7017-b764-452f-9783-c8ab0b90fbdc/volumes" Feb 20 08:45:09 crc kubenswrapper[4492]: I0220 08:45:09.311635 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:45:09 crc kubenswrapper[4492]: I0220 08:45:09.312183 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:45:10 crc kubenswrapper[4492]: I0220 08:45:10.133628 4492 scope.go:117] "RemoveContainer" containerID="c98c1a369d9eea98b825fbf139a60866c5cd77fc1de1d3eb8d469c8d58438d93" Feb 20 08:45:39 crc kubenswrapper[4492]: I0220 08:45:39.311130 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:45:39 crc kubenswrapper[4492]: I0220 08:45:39.311875 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:46:09 crc kubenswrapper[4492]: I0220 08:46:09.311083 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:46:09 crc kubenswrapper[4492]: I0220 08:46:09.311606 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:46:09 crc kubenswrapper[4492]: I0220 08:46:09.311657 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:46:09 crc kubenswrapper[4492]: I0220 08:46:09.312620 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ed761079a9284278df7d30a279bc0b5c625e1e30cacd794793b0f9fbe570624d"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:46:09 crc kubenswrapper[4492]: I0220 08:46:09.312683 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://ed761079a9284278df7d30a279bc0b5c625e1e30cacd794793b0f9fbe570624d" gracePeriod=600 Feb 20 08:46:10 crc kubenswrapper[4492]: I0220 08:46:10.274246 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="ed761079a9284278df7d30a279bc0b5c625e1e30cacd794793b0f9fbe570624d" exitCode=0 Feb 20 08:46:10 crc kubenswrapper[4492]: I0220 08:46:10.274329 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"ed761079a9284278df7d30a279bc0b5c625e1e30cacd794793b0f9fbe570624d"} Feb 20 08:46:10 crc kubenswrapper[4492]: I0220 08:46:10.275191 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc"} Feb 20 08:46:10 crc kubenswrapper[4492]: I0220 08:46:10.275291 4492 scope.go:117] "RemoveContainer" containerID="ae2886791501a3e1b0f07eb4f958bb86f6ba5cd13c887d3ded620b47bf288362" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.507411 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-79tnh"] Feb 20 08:48:05 crc kubenswrapper[4492]: E0220 08:48:05.508580 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d036832f-1f0c-4f37-be93-5173bc06e81f" containerName="collect-profiles" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.508600 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d036832f-1f0c-4f37-be93-5173bc06e81f" containerName="collect-profiles" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.508866 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d036832f-1f0c-4f37-be93-5173bc06e81f" containerName="collect-profiles" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.510244 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.528596 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-catalog-content\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.528876 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-utilities\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.529049 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blw5w\" (UniqueName: \"kubernetes.io/projected/f3532488-eb21-4198-9fe9-4226d8e0d8f8-kube-api-access-blw5w\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.536313 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-79tnh"] Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.630593 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-catalog-content\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.630933 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-utilities\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.631192 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blw5w\" (UniqueName: \"kubernetes.io/projected/f3532488-eb21-4198-9fe9-4226d8e0d8f8-kube-api-access-blw5w\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.633748 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-catalog-content\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.634517 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-utilities\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.654145 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blw5w\" (UniqueName: \"kubernetes.io/projected/f3532488-eb21-4198-9fe9-4226d8e0d8f8-kube-api-access-blw5w\") pod \"community-operators-79tnh\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:05 crc kubenswrapper[4492]: I0220 08:48:05.828962 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:06 crc kubenswrapper[4492]: I0220 08:48:06.447918 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-79tnh"] Feb 20 08:48:06 crc kubenswrapper[4492]: W0220 08:48:06.472012 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3532488_eb21_4198_9fe9_4226d8e0d8f8.slice/crio-950d91fa29d1448aa303c41cc5a30eca4c64e484888730082734fc2ad0c5730b WatchSource:0}: Error finding container 950d91fa29d1448aa303c41cc5a30eca4c64e484888730082734fc2ad0c5730b: Status 404 returned error can't find the container with id 950d91fa29d1448aa303c41cc5a30eca4c64e484888730082734fc2ad0c5730b Feb 20 08:48:07 crc kubenswrapper[4492]: I0220 08:48:07.228428 4492 generic.go:334] "Generic (PLEG): container finished" podID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerID="3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c" exitCode=0 Feb 20 08:48:07 crc kubenswrapper[4492]: I0220 08:48:07.228808 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79tnh" event={"ID":"f3532488-eb21-4198-9fe9-4226d8e0d8f8","Type":"ContainerDied","Data":"3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c"} Feb 20 08:48:07 crc kubenswrapper[4492]: I0220 08:48:07.228843 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79tnh" event={"ID":"f3532488-eb21-4198-9fe9-4226d8e0d8f8","Type":"ContainerStarted","Data":"950d91fa29d1448aa303c41cc5a30eca4c64e484888730082734fc2ad0c5730b"} Feb 20 08:48:07 crc kubenswrapper[4492]: I0220 08:48:07.233377 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:48:08 crc kubenswrapper[4492]: I0220 08:48:08.262912 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79tnh" event={"ID":"f3532488-eb21-4198-9fe9-4226d8e0d8f8","Type":"ContainerStarted","Data":"1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94"} Feb 20 08:48:09 crc kubenswrapper[4492]: I0220 08:48:09.287866 4492 generic.go:334] "Generic (PLEG): container finished" podID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerID="1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94" exitCode=0 Feb 20 08:48:09 crc kubenswrapper[4492]: I0220 08:48:09.288247 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79tnh" event={"ID":"f3532488-eb21-4198-9fe9-4226d8e0d8f8","Type":"ContainerDied","Data":"1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94"} Feb 20 08:48:09 crc kubenswrapper[4492]: I0220 08:48:09.312133 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:48:09 crc kubenswrapper[4492]: I0220 08:48:09.312213 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:48:10 crc kubenswrapper[4492]: I0220 08:48:10.298653 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79tnh" event={"ID":"f3532488-eb21-4198-9fe9-4226d8e0d8f8","Type":"ContainerStarted","Data":"83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692"} Feb 20 08:48:15 crc kubenswrapper[4492]: I0220 08:48:15.829912 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:15 crc kubenswrapper[4492]: I0220 08:48:15.830508 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:15 crc kubenswrapper[4492]: I0220 08:48:15.875755 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:15 crc kubenswrapper[4492]: I0220 08:48:15.897835 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-79tnh" podStartSLOduration=8.336697247 podStartE2EDuration="10.897814068s" podCreationTimestamp="2026-02-20 08:48:05 +0000 UTC" firstStartedPulling="2026-02-20 08:48:07.232063313 +0000 UTC m=+7644.003352291" lastFinishedPulling="2026-02-20 08:48:09.793180134 +0000 UTC m=+7646.564469112" observedRunningTime="2026-02-20 08:48:10.319170525 +0000 UTC m=+7647.090459503" watchObservedRunningTime="2026-02-20 08:48:15.897814068 +0000 UTC m=+7652.669103046" Feb 20 08:48:16 crc kubenswrapper[4492]: I0220 08:48:16.401525 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:16 crc kubenswrapper[4492]: I0220 08:48:16.446103 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-79tnh"] Feb 20 08:48:18 crc kubenswrapper[4492]: I0220 08:48:18.399916 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-79tnh" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="registry-server" containerID="cri-o://83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692" gracePeriod=2 Feb 20 08:48:18 crc kubenswrapper[4492]: I0220 08:48:18.909686 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.024958 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-catalog-content\") pod \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.025326 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blw5w\" (UniqueName: \"kubernetes.io/projected/f3532488-eb21-4198-9fe9-4226d8e0d8f8-kube-api-access-blw5w\") pod \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.025536 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-utilities\") pod \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\" (UID: \"f3532488-eb21-4198-9fe9-4226d8e0d8f8\") " Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.025905 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-utilities" (OuterVolumeSpecName: "utilities") pod "f3532488-eb21-4198-9fe9-4226d8e0d8f8" (UID: "f3532488-eb21-4198-9fe9-4226d8e0d8f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.026509 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.034655 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3532488-eb21-4198-9fe9-4226d8e0d8f8-kube-api-access-blw5w" (OuterVolumeSpecName: "kube-api-access-blw5w") pod "f3532488-eb21-4198-9fe9-4226d8e0d8f8" (UID: "f3532488-eb21-4198-9fe9-4226d8e0d8f8"). InnerVolumeSpecName "kube-api-access-blw5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.068709 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3532488-eb21-4198-9fe9-4226d8e0d8f8" (UID: "f3532488-eb21-4198-9fe9-4226d8e0d8f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.129370 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blw5w\" (UniqueName: \"kubernetes.io/projected/f3532488-eb21-4198-9fe9-4226d8e0d8f8-kube-api-access-blw5w\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.129406 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3532488-eb21-4198-9fe9-4226d8e0d8f8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.412886 4492 generic.go:334] "Generic (PLEG): container finished" podID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerID="83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692" exitCode=0 Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.412952 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79tnh" event={"ID":"f3532488-eb21-4198-9fe9-4226d8e0d8f8","Type":"ContainerDied","Data":"83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692"} Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.413022 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-79tnh" event={"ID":"f3532488-eb21-4198-9fe9-4226d8e0d8f8","Type":"ContainerDied","Data":"950d91fa29d1448aa303c41cc5a30eca4c64e484888730082734fc2ad0c5730b"} Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.413020 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-79tnh" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.413050 4492 scope.go:117] "RemoveContainer" containerID="83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.433105 4492 scope.go:117] "RemoveContainer" containerID="1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.444586 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-79tnh"] Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.452653 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-79tnh"] Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.476814 4492 scope.go:117] "RemoveContainer" containerID="3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.502627 4492 scope.go:117] "RemoveContainer" containerID="83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692" Feb 20 08:48:19 crc kubenswrapper[4492]: E0220 08:48:19.503104 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692\": container with ID starting with 83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692 not found: ID does not exist" containerID="83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.503150 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692"} err="failed to get container status \"83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692\": rpc error: code = NotFound desc = could not find container \"83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692\": container with ID starting with 83612a2b962e0fbd32e8b56c5d88f9825f5189ba3cb360b6265272a3ab351692 not found: ID does not exist" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.503186 4492 scope.go:117] "RemoveContainer" containerID="1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94" Feb 20 08:48:19 crc kubenswrapper[4492]: E0220 08:48:19.503768 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94\": container with ID starting with 1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94 not found: ID does not exist" containerID="1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.503892 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94"} err="failed to get container status \"1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94\": rpc error: code = NotFound desc = could not find container \"1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94\": container with ID starting with 1fa90c5bfb5b7b45392af6cee29a343ec53ff846c73cfce3f18d1aa6df725f94 not found: ID does not exist" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.503983 4492 scope.go:117] "RemoveContainer" containerID="3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c" Feb 20 08:48:19 crc kubenswrapper[4492]: E0220 08:48:19.504432 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c\": container with ID starting with 3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c not found: ID does not exist" containerID="3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.504462 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c"} err="failed to get container status \"3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c\": rpc error: code = NotFound desc = could not find container \"3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c\": container with ID starting with 3f47f451b4a160ae49260500922c2a5e839f15196b1b00771fcf8461a2763c9c not found: ID does not exist" Feb 20 08:48:19 crc kubenswrapper[4492]: I0220 08:48:19.568376 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" path="/var/lib/kubelet/pods/f3532488-eb21-4198-9fe9-4226d8e0d8f8/volumes" Feb 20 08:48:39 crc kubenswrapper[4492]: I0220 08:48:39.311748 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:48:39 crc kubenswrapper[4492]: I0220 08:48:39.312578 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.311648 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.312426 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.312520 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.313222 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.313291 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" gracePeriod=600 Feb 20 08:49:09 crc kubenswrapper[4492]: E0220 08:49:09.434410 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.861184 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" exitCode=0 Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.861273 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc"} Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.861567 4492 scope.go:117] "RemoveContainer" containerID="ed761079a9284278df7d30a279bc0b5c625e1e30cacd794793b0f9fbe570624d" Feb 20 08:49:09 crc kubenswrapper[4492]: I0220 08:49:09.862497 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:49:09 crc kubenswrapper[4492]: E0220 08:49:09.862762 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:49:23 crc kubenswrapper[4492]: I0220 08:49:23.565827 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:49:23 crc kubenswrapper[4492]: E0220 08:49:23.566950 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:49:35 crc kubenswrapper[4492]: I0220 08:49:35.557133 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:49:35 crc kubenswrapper[4492]: E0220 08:49:35.558027 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:49:46 crc kubenswrapper[4492]: I0220 08:49:46.558361 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:49:46 crc kubenswrapper[4492]: E0220 08:49:46.559701 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:49:58 crc kubenswrapper[4492]: I0220 08:49:58.556945 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:49:58 crc kubenswrapper[4492]: E0220 08:49:58.558083 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:50:13 crc kubenswrapper[4492]: I0220 08:50:13.562187 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:50:13 crc kubenswrapper[4492]: E0220 08:50:13.563036 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:50:27 crc kubenswrapper[4492]: I0220 08:50:27.558057 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:50:27 crc kubenswrapper[4492]: E0220 08:50:27.559153 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:50:41 crc kubenswrapper[4492]: I0220 08:50:41.557578 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:50:41 crc kubenswrapper[4492]: E0220 08:50:41.558696 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:50:52 crc kubenswrapper[4492]: I0220 08:50:52.557707 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:50:52 crc kubenswrapper[4492]: E0220 08:50:52.558993 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:51:07 crc kubenswrapper[4492]: I0220 08:51:07.561856 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:51:07 crc kubenswrapper[4492]: E0220 08:51:07.562719 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:51:19 crc kubenswrapper[4492]: I0220 08:51:19.557891 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:51:19 crc kubenswrapper[4492]: E0220 08:51:19.558748 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:51:33 crc kubenswrapper[4492]: I0220 08:51:33.563199 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:51:33 crc kubenswrapper[4492]: E0220 08:51:33.564540 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:51:46 crc kubenswrapper[4492]: I0220 08:51:46.558152 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:51:46 crc kubenswrapper[4492]: E0220 08:51:46.559431 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:51:58 crc kubenswrapper[4492]: I0220 08:51:58.557827 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:51:58 crc kubenswrapper[4492]: E0220 08:51:58.559051 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:52:13 crc kubenswrapper[4492]: I0220 08:52:13.563057 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:52:13 crc kubenswrapper[4492]: E0220 08:52:13.564802 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:52:24 crc kubenswrapper[4492]: I0220 08:52:24.556626 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:52:24 crc kubenswrapper[4492]: E0220 08:52:24.557454 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:52:36 crc kubenswrapper[4492]: I0220 08:52:36.556862 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:52:36 crc kubenswrapper[4492]: E0220 08:52:36.557771 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.153616 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fxj6k"] Feb 20 08:52:42 crc kubenswrapper[4492]: E0220 08:52:42.154841 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="registry-server" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.154866 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="registry-server" Feb 20 08:52:42 crc kubenswrapper[4492]: E0220 08:52:42.154895 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="extract-utilities" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.154902 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="extract-utilities" Feb 20 08:52:42 crc kubenswrapper[4492]: E0220 08:52:42.154926 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="extract-content" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.154932 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="extract-content" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.155138 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3532488-eb21-4198-9fe9-4226d8e0d8f8" containerName="registry-server" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.156641 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.174519 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxj6k"] Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.231550 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47fzp\" (UniqueName: \"kubernetes.io/projected/442a1ec3-652b-4b66-828a-b0e02f7f8387-kube-api-access-47fzp\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.231664 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-catalog-content\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.231781 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-utilities\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.333653 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47fzp\" (UniqueName: \"kubernetes.io/projected/442a1ec3-652b-4b66-828a-b0e02f7f8387-kube-api-access-47fzp\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.333724 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-catalog-content\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.333807 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-utilities\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.334217 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-utilities\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.334781 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-catalog-content\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.362973 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47fzp\" (UniqueName: \"kubernetes.io/projected/442a1ec3-652b-4b66-828a-b0e02f7f8387-kube-api-access-47fzp\") pod \"redhat-marketplace-fxj6k\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:42 crc kubenswrapper[4492]: I0220 08:52:42.480990 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:43 crc kubenswrapper[4492]: I0220 08:52:43.111300 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxj6k"] Feb 20 08:52:43 crc kubenswrapper[4492]: I0220 08:52:43.707766 4492 generic.go:334] "Generic (PLEG): container finished" podID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerID="9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0" exitCode=0 Feb 20 08:52:43 crc kubenswrapper[4492]: I0220 08:52:43.707823 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxj6k" event={"ID":"442a1ec3-652b-4b66-828a-b0e02f7f8387","Type":"ContainerDied","Data":"9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0"} Feb 20 08:52:43 crc kubenswrapper[4492]: I0220 08:52:43.708171 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxj6k" event={"ID":"442a1ec3-652b-4b66-828a-b0e02f7f8387","Type":"ContainerStarted","Data":"726d02a511da1cc4276f84cb0b72cf2973b034d4cef6da26a4b6d64443681ecf"} Feb 20 08:52:44 crc kubenswrapper[4492]: I0220 08:52:44.722291 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxj6k" event={"ID":"442a1ec3-652b-4b66-828a-b0e02f7f8387","Type":"ContainerStarted","Data":"f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057"} Feb 20 08:52:45 crc kubenswrapper[4492]: I0220 08:52:45.737098 4492 generic.go:334] "Generic (PLEG): container finished" podID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerID="f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057" exitCode=0 Feb 20 08:52:45 crc kubenswrapper[4492]: I0220 08:52:45.737300 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxj6k" event={"ID":"442a1ec3-652b-4b66-828a-b0e02f7f8387","Type":"ContainerDied","Data":"f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057"} Feb 20 08:52:46 crc kubenswrapper[4492]: I0220 08:52:46.753033 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxj6k" event={"ID":"442a1ec3-652b-4b66-828a-b0e02f7f8387","Type":"ContainerStarted","Data":"4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea"} Feb 20 08:52:46 crc kubenswrapper[4492]: I0220 08:52:46.783674 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fxj6k" podStartSLOduration=2.201523663 podStartE2EDuration="4.782627895s" podCreationTimestamp="2026-02-20 08:52:42 +0000 UTC" firstStartedPulling="2026-02-20 08:52:43.709805019 +0000 UTC m=+7920.481093986" lastFinishedPulling="2026-02-20 08:52:46.290909239 +0000 UTC m=+7923.062198218" observedRunningTime="2026-02-20 08:52:46.77830235 +0000 UTC m=+7923.549591328" watchObservedRunningTime="2026-02-20 08:52:46.782627895 +0000 UTC m=+7923.553916873" Feb 20 08:52:47 crc kubenswrapper[4492]: I0220 08:52:47.557306 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:52:47 crc kubenswrapper[4492]: E0220 08:52:47.558090 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:52:52 crc kubenswrapper[4492]: I0220 08:52:52.481146 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:52 crc kubenswrapper[4492]: I0220 08:52:52.481869 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:52 crc kubenswrapper[4492]: I0220 08:52:52.523985 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:52 crc kubenswrapper[4492]: I0220 08:52:52.846652 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:52 crc kubenswrapper[4492]: I0220 08:52:52.905236 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxj6k"] Feb 20 08:52:54 crc kubenswrapper[4492]: I0220 08:52:54.824797 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fxj6k" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="registry-server" containerID="cri-o://4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea" gracePeriod=2 Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.367773 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.558431 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-catalog-content\") pod \"442a1ec3-652b-4b66-828a-b0e02f7f8387\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.558598 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-utilities\") pod \"442a1ec3-652b-4b66-828a-b0e02f7f8387\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.558641 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47fzp\" (UniqueName: \"kubernetes.io/projected/442a1ec3-652b-4b66-828a-b0e02f7f8387-kube-api-access-47fzp\") pod \"442a1ec3-652b-4b66-828a-b0e02f7f8387\" (UID: \"442a1ec3-652b-4b66-828a-b0e02f7f8387\") " Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.559339 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-utilities" (OuterVolumeSpecName: "utilities") pod "442a1ec3-652b-4b66-828a-b0e02f7f8387" (UID: "442a1ec3-652b-4b66-828a-b0e02f7f8387"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.566350 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442a1ec3-652b-4b66-828a-b0e02f7f8387-kube-api-access-47fzp" (OuterVolumeSpecName: "kube-api-access-47fzp") pod "442a1ec3-652b-4b66-828a-b0e02f7f8387" (UID: "442a1ec3-652b-4b66-828a-b0e02f7f8387"). InnerVolumeSpecName "kube-api-access-47fzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.585605 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "442a1ec3-652b-4b66-828a-b0e02f7f8387" (UID: "442a1ec3-652b-4b66-828a-b0e02f7f8387"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.663261 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.663365 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/442a1ec3-652b-4b66-828a-b0e02f7f8387-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.663443 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47fzp\" (UniqueName: \"kubernetes.io/projected/442a1ec3-652b-4b66-828a-b0e02f7f8387-kube-api-access-47fzp\") on node \"crc\" DevicePath \"\"" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.835219 4492 generic.go:334] "Generic (PLEG): container finished" podID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerID="4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea" exitCode=0 Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.835297 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxj6k" event={"ID":"442a1ec3-652b-4b66-828a-b0e02f7f8387","Type":"ContainerDied","Data":"4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea"} Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.835372 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fxj6k" event={"ID":"442a1ec3-652b-4b66-828a-b0e02f7f8387","Type":"ContainerDied","Data":"726d02a511da1cc4276f84cb0b72cf2973b034d4cef6da26a4b6d64443681ecf"} Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.835396 4492 scope.go:117] "RemoveContainer" containerID="4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.835717 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fxj6k" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.859813 4492 scope.go:117] "RemoveContainer" containerID="f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.891156 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxj6k"] Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.898781 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fxj6k"] Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.908651 4492 scope.go:117] "RemoveContainer" containerID="9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.928547 4492 scope.go:117] "RemoveContainer" containerID="4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea" Feb 20 08:52:55 crc kubenswrapper[4492]: E0220 08:52:55.929312 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea\": container with ID starting with 4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea not found: ID does not exist" containerID="4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.929357 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea"} err="failed to get container status \"4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea\": rpc error: code = NotFound desc = could not find container \"4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea\": container with ID starting with 4257bd9d221e30b88ac8f2ece7ea12a89b947d113410099a6f81fc26afdee8ea not found: ID does not exist" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.929386 4492 scope.go:117] "RemoveContainer" containerID="f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057" Feb 20 08:52:55 crc kubenswrapper[4492]: E0220 08:52:55.929803 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057\": container with ID starting with f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057 not found: ID does not exist" containerID="f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.929834 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057"} err="failed to get container status \"f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057\": rpc error: code = NotFound desc = could not find container \"f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057\": container with ID starting with f5cb2be6b056444c3de588691e728ba942f3cd880a440d6298a477ac95260057 not found: ID does not exist" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.929857 4492 scope.go:117] "RemoveContainer" containerID="9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0" Feb 20 08:52:55 crc kubenswrapper[4492]: E0220 08:52:55.930138 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0\": container with ID starting with 9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0 not found: ID does not exist" containerID="9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0" Feb 20 08:52:55 crc kubenswrapper[4492]: I0220 08:52:55.930166 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0"} err="failed to get container status \"9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0\": rpc error: code = NotFound desc = could not find container \"9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0\": container with ID starting with 9665e9bd19b769e8cc9a01ebc6c52f73c6e4c7006904c8b167e8c5b997ff6bc0 not found: ID does not exist" Feb 20 08:52:57 crc kubenswrapper[4492]: I0220 08:52:57.569200 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" path="/var/lib/kubelet/pods/442a1ec3-652b-4b66-828a-b0e02f7f8387/volumes" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.462486 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6md7n"] Feb 20 08:53:01 crc kubenswrapper[4492]: E0220 08:53:01.463320 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="registry-server" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.463336 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="registry-server" Feb 20 08:53:01 crc kubenswrapper[4492]: E0220 08:53:01.463351 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="extract-utilities" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.463357 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="extract-utilities" Feb 20 08:53:01 crc kubenswrapper[4492]: E0220 08:53:01.463365 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="extract-content" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.463371 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="extract-content" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.463939 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="442a1ec3-652b-4b66-828a-b0e02f7f8387" containerName="registry-server" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.465199 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.478587 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6md7n"] Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.603048 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-utilities\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.603106 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c47qh\" (UniqueName: \"kubernetes.io/projected/a1c8a52b-cb57-4df8-970a-e97841fefcd0-kube-api-access-c47qh\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.603368 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-catalog-content\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.706341 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-catalog-content\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.706687 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-utilities\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.706744 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c47qh\" (UniqueName: \"kubernetes.io/projected/a1c8a52b-cb57-4df8-970a-e97841fefcd0-kube-api-access-c47qh\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.706854 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-catalog-content\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.707306 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-utilities\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.726112 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c47qh\" (UniqueName: \"kubernetes.io/projected/a1c8a52b-cb57-4df8-970a-e97841fefcd0-kube-api-access-c47qh\") pod \"redhat-operators-6md7n\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:01 crc kubenswrapper[4492]: I0220 08:53:01.792719 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:02 crc kubenswrapper[4492]: I0220 08:53:02.276024 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6md7n"] Feb 20 08:53:02 crc kubenswrapper[4492]: W0220 08:53:02.284126 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1c8a52b_cb57_4df8_970a_e97841fefcd0.slice/crio-eef120435896bac6c0785f0ad9de9784f7a1d1ea715a1384f9117fe9b567d8ff WatchSource:0}: Error finding container eef120435896bac6c0785f0ad9de9784f7a1d1ea715a1384f9117fe9b567d8ff: Status 404 returned error can't find the container with id eef120435896bac6c0785f0ad9de9784f7a1d1ea715a1384f9117fe9b567d8ff Feb 20 08:53:02 crc kubenswrapper[4492]: I0220 08:53:02.558842 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:53:02 crc kubenswrapper[4492]: E0220 08:53:02.559486 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:53:02 crc kubenswrapper[4492]: I0220 08:53:02.925891 4492 generic.go:334] "Generic (PLEG): container finished" podID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerID="335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5" exitCode=0 Feb 20 08:53:02 crc kubenswrapper[4492]: I0220 08:53:02.926251 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6md7n" event={"ID":"a1c8a52b-cb57-4df8-970a-e97841fefcd0","Type":"ContainerDied","Data":"335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5"} Feb 20 08:53:02 crc kubenswrapper[4492]: I0220 08:53:02.926293 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6md7n" event={"ID":"a1c8a52b-cb57-4df8-970a-e97841fefcd0","Type":"ContainerStarted","Data":"eef120435896bac6c0785f0ad9de9784f7a1d1ea715a1384f9117fe9b567d8ff"} Feb 20 08:53:03 crc kubenswrapper[4492]: I0220 08:53:03.951771 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6md7n" event={"ID":"a1c8a52b-cb57-4df8-970a-e97841fefcd0","Type":"ContainerStarted","Data":"47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63"} Feb 20 08:53:06 crc kubenswrapper[4492]: I0220 08:53:06.990236 4492 generic.go:334] "Generic (PLEG): container finished" podID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerID="47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63" exitCode=0 Feb 20 08:53:06 crc kubenswrapper[4492]: I0220 08:53:06.990596 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6md7n" event={"ID":"a1c8a52b-cb57-4df8-970a-e97841fefcd0","Type":"ContainerDied","Data":"47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63"} Feb 20 08:53:08 crc kubenswrapper[4492]: I0220 08:53:08.005215 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6md7n" event={"ID":"a1c8a52b-cb57-4df8-970a-e97841fefcd0","Type":"ContainerStarted","Data":"d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9"} Feb 20 08:53:08 crc kubenswrapper[4492]: I0220 08:53:08.026002 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6md7n" podStartSLOduration=2.47182444 podStartE2EDuration="7.025981516s" podCreationTimestamp="2026-02-20 08:53:01 +0000 UTC" firstStartedPulling="2026-02-20 08:53:02.928933797 +0000 UTC m=+7939.700222775" lastFinishedPulling="2026-02-20 08:53:07.483090882 +0000 UTC m=+7944.254379851" observedRunningTime="2026-02-20 08:53:08.024968086 +0000 UTC m=+7944.796257064" watchObservedRunningTime="2026-02-20 08:53:08.025981516 +0000 UTC m=+7944.797270484" Feb 20 08:53:11 crc kubenswrapper[4492]: I0220 08:53:11.792966 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:11 crc kubenswrapper[4492]: I0220 08:53:11.793682 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:12 crc kubenswrapper[4492]: I0220 08:53:12.831355 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6md7n" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="registry-server" probeResult="failure" output=< Feb 20 08:53:12 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:53:12 crc kubenswrapper[4492]: > Feb 20 08:53:13 crc kubenswrapper[4492]: I0220 08:53:13.564811 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:53:13 crc kubenswrapper[4492]: E0220 08:53:13.565161 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:53:21 crc kubenswrapper[4492]: I0220 08:53:21.832904 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:21 crc kubenswrapper[4492]: I0220 08:53:21.871181 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:22 crc kubenswrapper[4492]: I0220 08:53:22.067457 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6md7n"] Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.143024 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6md7n" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="registry-server" containerID="cri-o://d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9" gracePeriod=2 Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.594436 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.715518 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c47qh\" (UniqueName: \"kubernetes.io/projected/a1c8a52b-cb57-4df8-970a-e97841fefcd0-kube-api-access-c47qh\") pod \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.715685 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-catalog-content\") pod \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.715919 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-utilities\") pod \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\" (UID: \"a1c8a52b-cb57-4df8-970a-e97841fefcd0\") " Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.716651 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-utilities" (OuterVolumeSpecName: "utilities") pod "a1c8a52b-cb57-4df8-970a-e97841fefcd0" (UID: "a1c8a52b-cb57-4df8-970a-e97841fefcd0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.717332 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.723991 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c8a52b-cb57-4df8-970a-e97841fefcd0-kube-api-access-c47qh" (OuterVolumeSpecName: "kube-api-access-c47qh") pod "a1c8a52b-cb57-4df8-970a-e97841fefcd0" (UID: "a1c8a52b-cb57-4df8-970a-e97841fefcd0"). InnerVolumeSpecName "kube-api-access-c47qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.814132 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1c8a52b-cb57-4df8-970a-e97841fefcd0" (UID: "a1c8a52b-cb57-4df8-970a-e97841fefcd0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.821214 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c47qh\" (UniqueName: \"kubernetes.io/projected/a1c8a52b-cb57-4df8-970a-e97841fefcd0-kube-api-access-c47qh\") on node \"crc\" DevicePath \"\"" Feb 20 08:53:23 crc kubenswrapper[4492]: I0220 08:53:23.821248 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1c8a52b-cb57-4df8-970a-e97841fefcd0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.153704 4492 generic.go:334] "Generic (PLEG): container finished" podID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerID="d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9" exitCode=0 Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.153755 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6md7n" event={"ID":"a1c8a52b-cb57-4df8-970a-e97841fefcd0","Type":"ContainerDied","Data":"d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9"} Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.154120 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6md7n" event={"ID":"a1c8a52b-cb57-4df8-970a-e97841fefcd0","Type":"ContainerDied","Data":"eef120435896bac6c0785f0ad9de9784f7a1d1ea715a1384f9117fe9b567d8ff"} Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.154160 4492 scope.go:117] "RemoveContainer" containerID="d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.153816 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6md7n" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.174330 4492 scope.go:117] "RemoveContainer" containerID="47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.186840 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6md7n"] Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.194108 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6md7n"] Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.211034 4492 scope.go:117] "RemoveContainer" containerID="335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.234343 4492 scope.go:117] "RemoveContainer" containerID="d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9" Feb 20 08:53:24 crc kubenswrapper[4492]: E0220 08:53:24.234849 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9\": container with ID starting with d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9 not found: ID does not exist" containerID="d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.234896 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9"} err="failed to get container status \"d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9\": rpc error: code = NotFound desc = could not find container \"d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9\": container with ID starting with d0bde29426deb7bd6c042e9450e52a26b5ed8b4e47d2d3e6590cc314a35001f9 not found: ID does not exist" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.234924 4492 scope.go:117] "RemoveContainer" containerID="47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63" Feb 20 08:53:24 crc kubenswrapper[4492]: E0220 08:53:24.235250 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63\": container with ID starting with 47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63 not found: ID does not exist" containerID="47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.235284 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63"} err="failed to get container status \"47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63\": rpc error: code = NotFound desc = could not find container \"47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63\": container with ID starting with 47fdc0a2c579ee1c4d7cc1698cf6d533285619aea1c93b5eccbcc7c8f90d9c63 not found: ID does not exist" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.235314 4492 scope.go:117] "RemoveContainer" containerID="335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5" Feb 20 08:53:24 crc kubenswrapper[4492]: E0220 08:53:24.242759 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5\": container with ID starting with 335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5 not found: ID does not exist" containerID="335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.242874 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5"} err="failed to get container status \"335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5\": rpc error: code = NotFound desc = could not find container \"335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5\": container with ID starting with 335d0847da37126ba4ecf9325414b2e9cf44ab21c437b9d34cba51c6c53d20a5 not found: ID does not exist" Feb 20 08:53:24 crc kubenswrapper[4492]: I0220 08:53:24.556822 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:53:24 crc kubenswrapper[4492]: E0220 08:53:24.557158 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:53:25 crc kubenswrapper[4492]: I0220 08:53:25.567161 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" path="/var/lib/kubelet/pods/a1c8a52b-cb57-4df8-970a-e97841fefcd0/volumes" Feb 20 08:53:38 crc kubenswrapper[4492]: I0220 08:53:38.557557 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:53:38 crc kubenswrapper[4492]: E0220 08:53:38.558293 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:53:52 crc kubenswrapper[4492]: I0220 08:53:52.557515 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:53:52 crc kubenswrapper[4492]: E0220 08:53:52.558263 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:54:04 crc kubenswrapper[4492]: I0220 08:54:04.557792 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:54:04 crc kubenswrapper[4492]: E0220 08:54:04.560320 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 08:54:15 crc kubenswrapper[4492]: I0220 08:54:15.557185 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:54:16 crc kubenswrapper[4492]: I0220 08:54:16.606995 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"36defec4ce3e6e893d258737c62888427130fda25c48873dc4c172c8cfa53958"} Feb 20 08:56:39 crc kubenswrapper[4492]: I0220 08:56:39.311568 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:56:39 crc kubenswrapper[4492]: I0220 08:56:39.312602 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:57:09 crc kubenswrapper[4492]: I0220 08:57:09.311147 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:57:09 crc kubenswrapper[4492]: I0220 08:57:09.311885 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.310788 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.311641 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.311691 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.312407 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36defec4ce3e6e893d258737c62888427130fda25c48873dc4c172c8cfa53958"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.312460 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://36defec4ce3e6e893d258737c62888427130fda25c48873dc4c172c8cfa53958" gracePeriod=600 Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.572267 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="36defec4ce3e6e893d258737c62888427130fda25c48873dc4c172c8cfa53958" exitCode=0 Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.578625 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"36defec4ce3e6e893d258737c62888427130fda25c48873dc4c172c8cfa53958"} Feb 20 08:57:39 crc kubenswrapper[4492]: I0220 08:57:39.578689 4492 scope.go:117] "RemoveContainer" containerID="73e22bc74746b2fd83465e403f0e509adbc3cf297051a2b5ddce2512d0d193dc" Feb 20 08:57:40 crc kubenswrapper[4492]: I0220 08:57:40.584649 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d"} Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.955394 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d4db78989-6hrcq"] Feb 20 08:58:32 crc kubenswrapper[4492]: E0220 08:58:32.961377 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="extract-content" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.961570 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="extract-content" Feb 20 08:58:32 crc kubenswrapper[4492]: E0220 08:58:32.961676 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="registry-server" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.961762 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="registry-server" Feb 20 08:58:32 crc kubenswrapper[4492]: E0220 08:58:32.961856 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="extract-utilities" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.961928 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="extract-utilities" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.962579 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c8a52b-cb57-4df8-970a-e97841fefcd0" containerName="registry-server" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.966807 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.979400 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-httpd-config\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.979498 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-combined-ca-bundle\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.979543 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r5dp\" (UniqueName: \"kubernetes.io/projected/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-kube-api-access-7r5dp\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.979613 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-ovndb-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.979656 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-public-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.979971 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-internal-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:32 crc kubenswrapper[4492]: I0220 08:58:32.980338 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-config\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.012137 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d4db78989-6hrcq"] Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.083167 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-internal-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.083703 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-config\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.083916 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-httpd-config\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.084017 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-combined-ca-bundle\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.084104 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r5dp\" (UniqueName: \"kubernetes.io/projected/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-kube-api-access-7r5dp\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.084286 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-ovndb-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.084914 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-public-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.096819 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-ovndb-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.096813 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-public-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.098426 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-config\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.099442 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-internal-tls-certs\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.100597 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-httpd-config\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.109152 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r5dp\" (UniqueName: \"kubernetes.io/projected/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-kube-api-access-7r5dp\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.122624 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166-combined-ca-bundle\") pod \"neutron-5d4db78989-6hrcq\" (UID: \"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166\") " pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:33 crc kubenswrapper[4492]: I0220 08:58:33.292171 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:34 crc kubenswrapper[4492]: I0220 08:58:34.207397 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d4db78989-6hrcq"] Feb 20 08:58:34 crc kubenswrapper[4492]: W0220 08:58:34.223000 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddde1c2ae_ef84_46c0_b3e3_fd8aa23c2166.slice/crio-3ad350186a242b90d7fff20349365e44176f101990729cb6a23b199509f5df98 WatchSource:0}: Error finding container 3ad350186a242b90d7fff20349365e44176f101990729cb6a23b199509f5df98: Status 404 returned error can't find the container with id 3ad350186a242b90d7fff20349365e44176f101990729cb6a23b199509f5df98 Feb 20 08:58:35 crc kubenswrapper[4492]: I0220 08:58:35.146032 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d4db78989-6hrcq" event={"ID":"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166","Type":"ContainerStarted","Data":"27eb7d2cbe34f34a7260c26e3a8a60cd9169d51a1a6549787b68664b1e54ed12"} Feb 20 08:58:35 crc kubenswrapper[4492]: I0220 08:58:35.146107 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d4db78989-6hrcq" event={"ID":"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166","Type":"ContainerStarted","Data":"eb0ad27d651940bda0318868419ca2148c5ded4a1688c17373dab40c0a66c528"} Feb 20 08:58:35 crc kubenswrapper[4492]: I0220 08:58:35.146123 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d4db78989-6hrcq" event={"ID":"dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166","Type":"ContainerStarted","Data":"3ad350186a242b90d7fff20349365e44176f101990729cb6a23b199509f5df98"} Feb 20 08:58:35 crc kubenswrapper[4492]: I0220 08:58:35.146229 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:58:35 crc kubenswrapper[4492]: I0220 08:58:35.170947 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d4db78989-6hrcq" podStartSLOduration=3.170932678 podStartE2EDuration="3.170932678s" podCreationTimestamp="2026-02-20 08:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 08:58:35.167940566 +0000 UTC m=+8271.939229545" watchObservedRunningTime="2026-02-20 08:58:35.170932678 +0000 UTC m=+8271.942221657" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.154375 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bbg46"] Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.157434 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.185496 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bbg46"] Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.223502 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xtp7\" (UniqueName: \"kubernetes.io/projected/9245e7d3-4d4b-438f-8749-34e76163110d-kube-api-access-7xtp7\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.223623 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-utilities\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.223703 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-catalog-content\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.327113 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xtp7\" (UniqueName: \"kubernetes.io/projected/9245e7d3-4d4b-438f-8749-34e76163110d-kube-api-access-7xtp7\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.327263 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-utilities\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.327455 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-catalog-content\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.328182 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-utilities\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.328255 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-catalog-content\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.347161 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xtp7\" (UniqueName: \"kubernetes.io/projected/9245e7d3-4d4b-438f-8749-34e76163110d-kube-api-access-7xtp7\") pod \"community-operators-bbg46\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:42 crc kubenswrapper[4492]: I0220 08:58:42.479980 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:43 crc kubenswrapper[4492]: I0220 08:58:43.031906 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bbg46"] Feb 20 08:58:43 crc kubenswrapper[4492]: I0220 08:58:43.222454 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bbg46" event={"ID":"9245e7d3-4d4b-438f-8749-34e76163110d","Type":"ContainerStarted","Data":"976d1d07d6d3b39cec01e3e91ad184859472e9c8c43d23fcad6e27bafca6ad1d"} Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.235060 4492 generic.go:334] "Generic (PLEG): container finished" podID="9245e7d3-4d4b-438f-8749-34e76163110d" containerID="9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9" exitCode=0 Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.235363 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bbg46" event={"ID":"9245e7d3-4d4b-438f-8749-34e76163110d","Type":"ContainerDied","Data":"9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9"} Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.241212 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.349439 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mbxkr"] Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.351223 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.369196 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbxkr"] Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.497766 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8bnb\" (UniqueName: \"kubernetes.io/projected/f5f29738-2e29-483a-a468-5aa137204317-kube-api-access-t8bnb\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.497820 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-utilities\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.497890 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-catalog-content\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.600628 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8bnb\" (UniqueName: \"kubernetes.io/projected/f5f29738-2e29-483a-a468-5aa137204317-kube-api-access-t8bnb\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.600781 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-utilities\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.601002 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-catalog-content\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.602321 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-utilities\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.602624 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-catalog-content\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.624238 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8bnb\" (UniqueName: \"kubernetes.io/projected/f5f29738-2e29-483a-a468-5aa137204317-kube-api-access-t8bnb\") pod \"certified-operators-mbxkr\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:44 crc kubenswrapper[4492]: I0220 08:58:44.666799 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:45 crc kubenswrapper[4492]: I0220 08:58:45.188966 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbxkr"] Feb 20 08:58:45 crc kubenswrapper[4492]: I0220 08:58:45.247202 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bbg46" event={"ID":"9245e7d3-4d4b-438f-8749-34e76163110d","Type":"ContainerStarted","Data":"f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968"} Feb 20 08:58:45 crc kubenswrapper[4492]: I0220 08:58:45.253774 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbxkr" event={"ID":"f5f29738-2e29-483a-a468-5aa137204317","Type":"ContainerStarted","Data":"03c7863bb814badb258a73f4d5fa4383e502a2d1cbcc98e79d07b00672b925ab"} Feb 20 08:58:46 crc kubenswrapper[4492]: I0220 08:58:46.263850 4492 generic.go:334] "Generic (PLEG): container finished" podID="9245e7d3-4d4b-438f-8749-34e76163110d" containerID="f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968" exitCode=0 Feb 20 08:58:46 crc kubenswrapper[4492]: I0220 08:58:46.263925 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bbg46" event={"ID":"9245e7d3-4d4b-438f-8749-34e76163110d","Type":"ContainerDied","Data":"f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968"} Feb 20 08:58:46 crc kubenswrapper[4492]: I0220 08:58:46.268491 4492 generic.go:334] "Generic (PLEG): container finished" podID="f5f29738-2e29-483a-a468-5aa137204317" containerID="0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a" exitCode=0 Feb 20 08:58:46 crc kubenswrapper[4492]: I0220 08:58:46.268560 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbxkr" event={"ID":"f5f29738-2e29-483a-a468-5aa137204317","Type":"ContainerDied","Data":"0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a"} Feb 20 08:58:47 crc kubenswrapper[4492]: I0220 08:58:47.284416 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bbg46" event={"ID":"9245e7d3-4d4b-438f-8749-34e76163110d","Type":"ContainerStarted","Data":"afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7"} Feb 20 08:58:47 crc kubenswrapper[4492]: I0220 08:58:47.287456 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbxkr" event={"ID":"f5f29738-2e29-483a-a468-5aa137204317","Type":"ContainerStarted","Data":"2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81"} Feb 20 08:58:47 crc kubenswrapper[4492]: I0220 08:58:47.314286 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bbg46" podStartSLOduration=2.822688689 podStartE2EDuration="5.314271334s" podCreationTimestamp="2026-02-20 08:58:42 +0000 UTC" firstStartedPulling="2026-02-20 08:58:44.237916477 +0000 UTC m=+8281.009205455" lastFinishedPulling="2026-02-20 08:58:46.729499122 +0000 UTC m=+8283.500788100" observedRunningTime="2026-02-20 08:58:47.303722383 +0000 UTC m=+8284.075011362" watchObservedRunningTime="2026-02-20 08:58:47.314271334 +0000 UTC m=+8284.085560313" Feb 20 08:58:48 crc kubenswrapper[4492]: I0220 08:58:48.303509 4492 generic.go:334] "Generic (PLEG): container finished" podID="f5f29738-2e29-483a-a468-5aa137204317" containerID="2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81" exitCode=0 Feb 20 08:58:48 crc kubenswrapper[4492]: I0220 08:58:48.303623 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbxkr" event={"ID":"f5f29738-2e29-483a-a468-5aa137204317","Type":"ContainerDied","Data":"2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81"} Feb 20 08:58:49 crc kubenswrapper[4492]: I0220 08:58:49.318296 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbxkr" event={"ID":"f5f29738-2e29-483a-a468-5aa137204317","Type":"ContainerStarted","Data":"b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf"} Feb 20 08:58:49 crc kubenswrapper[4492]: I0220 08:58:49.351216 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mbxkr" podStartSLOduration=2.846878156 podStartE2EDuration="5.351190421s" podCreationTimestamp="2026-02-20 08:58:44 +0000 UTC" firstStartedPulling="2026-02-20 08:58:46.271758001 +0000 UTC m=+8283.043046980" lastFinishedPulling="2026-02-20 08:58:48.776070257 +0000 UTC m=+8285.547359245" observedRunningTime="2026-02-20 08:58:49.346617188 +0000 UTC m=+8286.117906166" watchObservedRunningTime="2026-02-20 08:58:49.351190421 +0000 UTC m=+8286.122479399" Feb 20 08:58:52 crc kubenswrapper[4492]: I0220 08:58:52.480443 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:52 crc kubenswrapper[4492]: I0220 08:58:52.494574 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:58:53 crc kubenswrapper[4492]: I0220 08:58:53.587376 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-bbg46" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="registry-server" probeResult="failure" output=< Feb 20 08:58:53 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:58:53 crc kubenswrapper[4492]: > Feb 20 08:58:54 crc kubenswrapper[4492]: I0220 08:58:54.667545 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:54 crc kubenswrapper[4492]: I0220 08:58:54.667954 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:58:55 crc kubenswrapper[4492]: I0220 08:58:55.715033 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mbxkr" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="registry-server" probeResult="failure" output=< Feb 20 08:58:55 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 08:58:55 crc kubenswrapper[4492]: > Feb 20 08:59:02 crc kubenswrapper[4492]: I0220 08:59:02.524414 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:59:02 crc kubenswrapper[4492]: I0220 08:59:02.577791 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:59:02 crc kubenswrapper[4492]: I0220 08:59:02.777986 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bbg46"] Feb 20 08:59:03 crc kubenswrapper[4492]: I0220 08:59:03.320019 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d4db78989-6hrcq" Feb 20 08:59:03 crc kubenswrapper[4492]: I0220 08:59:03.510670 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-68fcdb65f7-pbqcq"] Feb 20 08:59:03 crc kubenswrapper[4492]: I0220 08:59:03.517488 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-68fcdb65f7-pbqcq" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-httpd" containerID="cri-o://55fc28fed46a899b668032714898309dd768815e5a9f5eb431a67a6f1574ce03" gracePeriod=30 Feb 20 08:59:03 crc kubenswrapper[4492]: I0220 08:59:03.517449 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-68fcdb65f7-pbqcq" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-api" containerID="cri-o://3f56ae764876ab87e13df3f6dd4387d92194dea3e5856e1f77114c7efee5a7ec" gracePeriod=30 Feb 20 08:59:04 crc kubenswrapper[4492]: I0220 08:59:04.483722 4492 generic.go:334] "Generic (PLEG): container finished" podID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerID="55fc28fed46a899b668032714898309dd768815e5a9f5eb431a67a6f1574ce03" exitCode=0 Feb 20 08:59:04 crc kubenswrapper[4492]: I0220 08:59:04.484325 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bbg46" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="registry-server" containerID="cri-o://afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7" gracePeriod=2 Feb 20 08:59:04 crc kubenswrapper[4492]: I0220 08:59:04.483931 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68fcdb65f7-pbqcq" event={"ID":"3f540de0-30e4-4b8b-b8de-628f4d46aa4c","Type":"ContainerDied","Data":"55fc28fed46a899b668032714898309dd768815e5a9f5eb431a67a6f1574ce03"} Feb 20 08:59:04 crc kubenswrapper[4492]: I0220 08:59:04.716196 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:59:04 crc kubenswrapper[4492]: I0220 08:59:04.770292 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.177491 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbxkr"] Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.294022 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.414085 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xtp7\" (UniqueName: \"kubernetes.io/projected/9245e7d3-4d4b-438f-8749-34e76163110d-kube-api-access-7xtp7\") pod \"9245e7d3-4d4b-438f-8749-34e76163110d\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.414191 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-utilities\") pod \"9245e7d3-4d4b-438f-8749-34e76163110d\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.414362 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-catalog-content\") pod \"9245e7d3-4d4b-438f-8749-34e76163110d\" (UID: \"9245e7d3-4d4b-438f-8749-34e76163110d\") " Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.415519 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-utilities" (OuterVolumeSpecName: "utilities") pod "9245e7d3-4d4b-438f-8749-34e76163110d" (UID: "9245e7d3-4d4b-438f-8749-34e76163110d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.436187 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9245e7d3-4d4b-438f-8749-34e76163110d-kube-api-access-7xtp7" (OuterVolumeSpecName: "kube-api-access-7xtp7") pod "9245e7d3-4d4b-438f-8749-34e76163110d" (UID: "9245e7d3-4d4b-438f-8749-34e76163110d"). InnerVolumeSpecName "kube-api-access-7xtp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.469032 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9245e7d3-4d4b-438f-8749-34e76163110d" (UID: "9245e7d3-4d4b-438f-8749-34e76163110d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.495131 4492 generic.go:334] "Generic (PLEG): container finished" podID="9245e7d3-4d4b-438f-8749-34e76163110d" containerID="afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7" exitCode=0 Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.495202 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bbg46" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.495292 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bbg46" event={"ID":"9245e7d3-4d4b-438f-8749-34e76163110d","Type":"ContainerDied","Data":"afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7"} Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.495325 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bbg46" event={"ID":"9245e7d3-4d4b-438f-8749-34e76163110d","Type":"ContainerDied","Data":"976d1d07d6d3b39cec01e3e91ad184859472e9c8c43d23fcad6e27bafca6ad1d"} Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.495347 4492 scope.go:117] "RemoveContainer" containerID="afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.517281 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xtp7\" (UniqueName: \"kubernetes.io/projected/9245e7d3-4d4b-438f-8749-34e76163110d-kube-api-access-7xtp7\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.517311 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.517321 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9245e7d3-4d4b-438f-8749-34e76163110d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.532214 4492 scope.go:117] "RemoveContainer" containerID="f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.536276 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bbg46"] Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.544605 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bbg46"] Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.552179 4492 scope.go:117] "RemoveContainer" containerID="9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.568789 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" path="/var/lib/kubelet/pods/9245e7d3-4d4b-438f-8749-34e76163110d/volumes" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.593865 4492 scope.go:117] "RemoveContainer" containerID="afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7" Feb 20 08:59:05 crc kubenswrapper[4492]: E0220 08:59:05.598305 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7\": container with ID starting with afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7 not found: ID does not exist" containerID="afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.598362 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7"} err="failed to get container status \"afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7\": rpc error: code = NotFound desc = could not find container \"afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7\": container with ID starting with afa387a819c536209169ac02883d63c3faf30989a7ae8a0ae4b9370ff06d08d7 not found: ID does not exist" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.598385 4492 scope.go:117] "RemoveContainer" containerID="f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968" Feb 20 08:59:05 crc kubenswrapper[4492]: E0220 08:59:05.598754 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968\": container with ID starting with f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968 not found: ID does not exist" containerID="f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.598780 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968"} err="failed to get container status \"f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968\": rpc error: code = NotFound desc = could not find container \"f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968\": container with ID starting with f05aa3265ded0cb4518a5f63bfa634449e85acbc72eb54849391d89eaa2b8968 not found: ID does not exist" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.598794 4492 scope.go:117] "RemoveContainer" containerID="9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9" Feb 20 08:59:05 crc kubenswrapper[4492]: E0220 08:59:05.599259 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9\": container with ID starting with 9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9 not found: ID does not exist" containerID="9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9" Feb 20 08:59:05 crc kubenswrapper[4492]: I0220 08:59:05.599280 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9"} err="failed to get container status \"9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9\": rpc error: code = NotFound desc = could not find container \"9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9\": container with ID starting with 9fa18e383bb33e63ae5e0a1510876f648430c6a756ec74adf2581898f1daa6c9 not found: ID does not exist" Feb 20 08:59:06 crc kubenswrapper[4492]: I0220 08:59:06.503452 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mbxkr" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="registry-server" containerID="cri-o://b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf" gracePeriod=2 Feb 20 08:59:06 crc kubenswrapper[4492]: I0220 08:59:06.934456 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.048181 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-utilities\") pod \"f5f29738-2e29-483a-a468-5aa137204317\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.048412 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-catalog-content\") pod \"f5f29738-2e29-483a-a468-5aa137204317\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.048523 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8bnb\" (UniqueName: \"kubernetes.io/projected/f5f29738-2e29-483a-a468-5aa137204317-kube-api-access-t8bnb\") pod \"f5f29738-2e29-483a-a468-5aa137204317\" (UID: \"f5f29738-2e29-483a-a468-5aa137204317\") " Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.049135 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-utilities" (OuterVolumeSpecName: "utilities") pod "f5f29738-2e29-483a-a468-5aa137204317" (UID: "f5f29738-2e29-483a-a468-5aa137204317"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.049304 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.060492 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5f29738-2e29-483a-a468-5aa137204317-kube-api-access-t8bnb" (OuterVolumeSpecName: "kube-api-access-t8bnb") pod "f5f29738-2e29-483a-a468-5aa137204317" (UID: "f5f29738-2e29-483a-a468-5aa137204317"). InnerVolumeSpecName "kube-api-access-t8bnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.094509 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5f29738-2e29-483a-a468-5aa137204317" (UID: "f5f29738-2e29-483a-a468-5aa137204317"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.152785 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5f29738-2e29-483a-a468-5aa137204317-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.152831 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8bnb\" (UniqueName: \"kubernetes.io/projected/f5f29738-2e29-483a-a468-5aa137204317-kube-api-access-t8bnb\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.522168 4492 generic.go:334] "Generic (PLEG): container finished" podID="f5f29738-2e29-483a-a468-5aa137204317" containerID="b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf" exitCode=0 Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.522227 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbxkr" event={"ID":"f5f29738-2e29-483a-a468-5aa137204317","Type":"ContainerDied","Data":"b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf"} Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.522260 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbxkr" event={"ID":"f5f29738-2e29-483a-a468-5aa137204317","Type":"ContainerDied","Data":"03c7863bb814badb258a73f4d5fa4383e502a2d1cbcc98e79d07b00672b925ab"} Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.522281 4492 scope.go:117] "RemoveContainer" containerID="b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.522422 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbxkr" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.553623 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbxkr"] Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.560261 4492 scope.go:117] "RemoveContainer" containerID="2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.572744 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mbxkr"] Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.607938 4492 scope.go:117] "RemoveContainer" containerID="0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.641972 4492 scope.go:117] "RemoveContainer" containerID="b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf" Feb 20 08:59:07 crc kubenswrapper[4492]: E0220 08:59:07.643048 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf\": container with ID starting with b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf not found: ID does not exist" containerID="b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.643108 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf"} err="failed to get container status \"b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf\": rpc error: code = NotFound desc = could not find container \"b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf\": container with ID starting with b343c42098acebd99081c3c78a46555742df0fd0234eb8bc5df6783e85757ecf not found: ID does not exist" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.643139 4492 scope.go:117] "RemoveContainer" containerID="2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81" Feb 20 08:59:07 crc kubenswrapper[4492]: E0220 08:59:07.643976 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81\": container with ID starting with 2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81 not found: ID does not exist" containerID="2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.644024 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81"} err="failed to get container status \"2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81\": rpc error: code = NotFound desc = could not find container \"2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81\": container with ID starting with 2d2ffc1b2e134775aa6ea1851ab7325341211a76738eda7d0f464c9bf980ac81 not found: ID does not exist" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.644053 4492 scope.go:117] "RemoveContainer" containerID="0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a" Feb 20 08:59:07 crc kubenswrapper[4492]: E0220 08:59:07.644610 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a\": container with ID starting with 0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a not found: ID does not exist" containerID="0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a" Feb 20 08:59:07 crc kubenswrapper[4492]: I0220 08:59:07.644638 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a"} err="failed to get container status \"0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a\": rpc error: code = NotFound desc = could not find container \"0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a\": container with ID starting with 0ab2023aea6a2bef7e10575d27b6d8f713138fd566763f6eba74d739c2dd364a not found: ID does not exist" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.577617 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5f29738-2e29-483a-a468-5aa137204317" path="/var/lib/kubelet/pods/f5f29738-2e29-483a-a468-5aa137204317/volumes" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.581195 4492 generic.go:334] "Generic (PLEG): container finished" podID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerID="3f56ae764876ab87e13df3f6dd4387d92194dea3e5856e1f77114c7efee5a7ec" exitCode=0 Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.581265 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68fcdb65f7-pbqcq" event={"ID":"3f540de0-30e4-4b8b-b8de-628f4d46aa4c","Type":"ContainerDied","Data":"3f56ae764876ab87e13df3f6dd4387d92194dea3e5856e1f77114c7efee5a7ec"} Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.725285 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.820493 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-public-tls-certs\") pod \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.820960 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w68zl\" (UniqueName: \"kubernetes.io/projected/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-kube-api-access-w68zl\") pod \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.821209 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-ovndb-tls-certs\") pod \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.821522 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-internal-tls-certs\") pod \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.821655 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-combined-ca-bundle\") pod \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.821825 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-httpd-config\") pod \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.821849 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-config\") pod \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\" (UID: \"3f540de0-30e4-4b8b-b8de-628f4d46aa4c\") " Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.829549 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-kube-api-access-w68zl" (OuterVolumeSpecName: "kube-api-access-w68zl") pod "3f540de0-30e4-4b8b-b8de-628f4d46aa4c" (UID: "3f540de0-30e4-4b8b-b8de-628f4d46aa4c"). InnerVolumeSpecName "kube-api-access-w68zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.841389 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "3f540de0-30e4-4b8b-b8de-628f4d46aa4c" (UID: "3f540de0-30e4-4b8b-b8de-628f4d46aa4c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.875007 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3f540de0-30e4-4b8b-b8de-628f4d46aa4c" (UID: "3f540de0-30e4-4b8b-b8de-628f4d46aa4c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.878493 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3f540de0-30e4-4b8b-b8de-628f4d46aa4c" (UID: "3f540de0-30e4-4b8b-b8de-628f4d46aa4c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.880952 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-config" (OuterVolumeSpecName: "config") pod "3f540de0-30e4-4b8b-b8de-628f4d46aa4c" (UID: "3f540de0-30e4-4b8b-b8de-628f4d46aa4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.889573 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f540de0-30e4-4b8b-b8de-628f4d46aa4c" (UID: "3f540de0-30e4-4b8b-b8de-628f4d46aa4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.915272 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "3f540de0-30e4-4b8b-b8de-628f4d46aa4c" (UID: "3f540de0-30e4-4b8b-b8de-628f4d46aa4c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.925721 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w68zl\" (UniqueName: \"kubernetes.io/projected/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-kube-api-access-w68zl\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.925763 4492 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.925773 4492 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.925782 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.925792 4492 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.925803 4492 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-config\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:09 crc kubenswrapper[4492]: I0220 08:59:09.925814 4492 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f540de0-30e4-4b8b-b8de-628f4d46aa4c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 20 08:59:10 crc kubenswrapper[4492]: I0220 08:59:10.486840 4492 scope.go:117] "RemoveContainer" containerID="3f56ae764876ab87e13df3f6dd4387d92194dea3e5856e1f77114c7efee5a7ec" Feb 20 08:59:10 crc kubenswrapper[4492]: I0220 08:59:10.511107 4492 scope.go:117] "RemoveContainer" containerID="55fc28fed46a899b668032714898309dd768815e5a9f5eb431a67a6f1574ce03" Feb 20 08:59:10 crc kubenswrapper[4492]: I0220 08:59:10.593776 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-68fcdb65f7-pbqcq" Feb 20 08:59:10 crc kubenswrapper[4492]: I0220 08:59:10.593879 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-68fcdb65f7-pbqcq" event={"ID":"3f540de0-30e4-4b8b-b8de-628f4d46aa4c","Type":"ContainerDied","Data":"071741118bfdd7557b22826e3e2d2ccf495bc3fb30a14cfa81a16a7da77912f4"} Feb 20 08:59:10 crc kubenswrapper[4492]: I0220 08:59:10.651289 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-68fcdb65f7-pbqcq"] Feb 20 08:59:10 crc kubenswrapper[4492]: I0220 08:59:10.663132 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-68fcdb65f7-pbqcq"] Feb 20 08:59:11 crc kubenswrapper[4492]: I0220 08:59:11.569438 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" path="/var/lib/kubelet/pods/3f540de0-30e4-4b8b-b8de-628f4d46aa4c/volumes" Feb 20 08:59:39 crc kubenswrapper[4492]: I0220 08:59:39.311419 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 08:59:39 crc kubenswrapper[4492]: I0220 08:59:39.312578 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.243858 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls"] Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.246148 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-api" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.246803 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-api" Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.246897 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-httpd" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.246954 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-httpd" Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.247016 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="registry-server" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.247060 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="registry-server" Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.247115 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="extract-content" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.247160 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="extract-content" Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.247206 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="extract-utilities" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.247252 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="extract-utilities" Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.247298 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="registry-server" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.247339 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="registry-server" Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.247388 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="extract-content" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.247431 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="extract-content" Feb 20 09:00:00 crc kubenswrapper[4492]: E0220 09:00:00.247506 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="extract-utilities" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.247591 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="extract-utilities" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.251261 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5f29738-2e29-483a-a468-5aa137204317" containerName="registry-server" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.251395 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-api" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.251453 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f540de0-30e4-4b8b-b8de-628f4d46aa4c" containerName="neutron-httpd" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.251536 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="9245e7d3-4d4b-438f-8749-34e76163110d" containerName="registry-server" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.253220 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.259605 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.260165 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.264071 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls"] Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.423692 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc4rq\" (UniqueName: \"kubernetes.io/projected/0599b1ae-e434-4dae-921c-86830dd8bec7-kube-api-access-hc4rq\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.424553 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0599b1ae-e434-4dae-921c-86830dd8bec7-config-volume\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.424770 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0599b1ae-e434-4dae-921c-86830dd8bec7-secret-volume\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.528464 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc4rq\" (UniqueName: \"kubernetes.io/projected/0599b1ae-e434-4dae-921c-86830dd8bec7-kube-api-access-hc4rq\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.528562 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0599b1ae-e434-4dae-921c-86830dd8bec7-config-volume\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.528663 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0599b1ae-e434-4dae-921c-86830dd8bec7-secret-volume\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.529694 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0599b1ae-e434-4dae-921c-86830dd8bec7-config-volume\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.543284 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0599b1ae-e434-4dae-921c-86830dd8bec7-secret-volume\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.546662 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc4rq\" (UniqueName: \"kubernetes.io/projected/0599b1ae-e434-4dae-921c-86830dd8bec7-kube-api-access-hc4rq\") pod \"collect-profiles-29526300-c79ls\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:00 crc kubenswrapper[4492]: I0220 09:00:00.577519 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:01 crc kubenswrapper[4492]: I0220 09:00:01.044598 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls"] Feb 20 09:00:01 crc kubenswrapper[4492]: I0220 09:00:01.102850 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" event={"ID":"0599b1ae-e434-4dae-921c-86830dd8bec7","Type":"ContainerStarted","Data":"c7169c7b5c6d7ef076c13278f3948ef75f9c411a0f67b77a7ccf409f23816898"} Feb 20 09:00:02 crc kubenswrapper[4492]: I0220 09:00:02.112885 4492 generic.go:334] "Generic (PLEG): container finished" podID="0599b1ae-e434-4dae-921c-86830dd8bec7" containerID="006cd035dd923e12a6c603f1953d8d7af2b473f317807f1243a07357b2a1246d" exitCode=0 Feb 20 09:00:02 crc kubenswrapper[4492]: I0220 09:00:02.113338 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" event={"ID":"0599b1ae-e434-4dae-921c-86830dd8bec7","Type":"ContainerDied","Data":"006cd035dd923e12a6c603f1953d8d7af2b473f317807f1243a07357b2a1246d"} Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.462586 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.607924 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc4rq\" (UniqueName: \"kubernetes.io/projected/0599b1ae-e434-4dae-921c-86830dd8bec7-kube-api-access-hc4rq\") pod \"0599b1ae-e434-4dae-921c-86830dd8bec7\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.608031 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0599b1ae-e434-4dae-921c-86830dd8bec7-secret-volume\") pod \"0599b1ae-e434-4dae-921c-86830dd8bec7\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.608183 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0599b1ae-e434-4dae-921c-86830dd8bec7-config-volume\") pod \"0599b1ae-e434-4dae-921c-86830dd8bec7\" (UID: \"0599b1ae-e434-4dae-921c-86830dd8bec7\") " Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.608904 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0599b1ae-e434-4dae-921c-86830dd8bec7-config-volume" (OuterVolumeSpecName: "config-volume") pod "0599b1ae-e434-4dae-921c-86830dd8bec7" (UID: "0599b1ae-e434-4dae-921c-86830dd8bec7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.615512 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0599b1ae-e434-4dae-921c-86830dd8bec7-kube-api-access-hc4rq" (OuterVolumeSpecName: "kube-api-access-hc4rq") pod "0599b1ae-e434-4dae-921c-86830dd8bec7" (UID: "0599b1ae-e434-4dae-921c-86830dd8bec7"). InnerVolumeSpecName "kube-api-access-hc4rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.616723 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0599b1ae-e434-4dae-921c-86830dd8bec7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0599b1ae-e434-4dae-921c-86830dd8bec7" (UID: "0599b1ae-e434-4dae-921c-86830dd8bec7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.709942 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0599b1ae-e434-4dae-921c-86830dd8bec7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.709983 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hc4rq\" (UniqueName: \"kubernetes.io/projected/0599b1ae-e434-4dae-921c-86830dd8bec7-kube-api-access-hc4rq\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:03 crc kubenswrapper[4492]: I0220 09:00:03.709996 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0599b1ae-e434-4dae-921c-86830dd8bec7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:00:04 crc kubenswrapper[4492]: I0220 09:00:04.132932 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" event={"ID":"0599b1ae-e434-4dae-921c-86830dd8bec7","Type":"ContainerDied","Data":"c7169c7b5c6d7ef076c13278f3948ef75f9c411a0f67b77a7ccf409f23816898"} Feb 20 09:00:04 crc kubenswrapper[4492]: I0220 09:00:04.133288 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7169c7b5c6d7ef076c13278f3948ef75f9c411a0f67b77a7ccf409f23816898" Feb 20 09:00:04 crc kubenswrapper[4492]: I0220 09:00:04.133227 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526300-c79ls" Feb 20 09:00:04 crc kubenswrapper[4492]: I0220 09:00:04.545360 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv"] Feb 20 09:00:04 crc kubenswrapper[4492]: I0220 09:00:04.555120 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526255-8tsjv"] Feb 20 09:00:05 crc kubenswrapper[4492]: I0220 09:00:05.569399 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64892d2a-51dd-4ae6-8936-3c9591b32884" path="/var/lib/kubelet/pods/64892d2a-51dd-4ae6-8936-3c9591b32884/volumes" Feb 20 09:00:09 crc kubenswrapper[4492]: I0220 09:00:09.311776 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:00:09 crc kubenswrapper[4492]: I0220 09:00:09.312625 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:00:10 crc kubenswrapper[4492]: I0220 09:00:10.652286 4492 scope.go:117] "RemoveContainer" containerID="b18eee7fda7f7686f5eecca57bf90554594eb9c243edfdf9ac19c79598cb5a8e" Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.311154 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.311986 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.312048 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.313131 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.313205 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" gracePeriod=600 Feb 20 09:00:39 crc kubenswrapper[4492]: E0220 09:00:39.430270 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.470280 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" exitCode=0 Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.470336 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d"} Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.470416 4492 scope.go:117] "RemoveContainer" containerID="36defec4ce3e6e893d258737c62888427130fda25c48873dc4c172c8cfa53958" Feb 20 09:00:39 crc kubenswrapper[4492]: I0220 09:00:39.471527 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:00:39 crc kubenswrapper[4492]: E0220 09:00:39.471952 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:00:54 crc kubenswrapper[4492]: I0220 09:00:54.557590 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:00:54 crc kubenswrapper[4492]: E0220 09:00:54.558282 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.150140 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29526301-gkh5l"] Feb 20 09:01:00 crc kubenswrapper[4492]: E0220 09:01:00.151185 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0599b1ae-e434-4dae-921c-86830dd8bec7" containerName="collect-profiles" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.151200 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0599b1ae-e434-4dae-921c-86830dd8bec7" containerName="collect-profiles" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.151384 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0599b1ae-e434-4dae-921c-86830dd8bec7" containerName="collect-profiles" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.152031 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.169568 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526301-gkh5l"] Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.261413 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-config-data\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.261491 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4qn6\" (UniqueName: \"kubernetes.io/projected/c66090c8-b2eb-45ae-9504-2d11f6cd1819-kube-api-access-j4qn6\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.261528 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-fernet-keys\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.261575 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-combined-ca-bundle\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.363586 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-combined-ca-bundle\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.363726 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-config-data\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.363772 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4qn6\" (UniqueName: \"kubernetes.io/projected/c66090c8-b2eb-45ae-9504-2d11f6cd1819-kube-api-access-j4qn6\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.363803 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-fernet-keys\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.371594 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-fernet-keys\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.372499 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-config-data\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.372632 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-combined-ca-bundle\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.380699 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4qn6\" (UniqueName: \"kubernetes.io/projected/c66090c8-b2eb-45ae-9504-2d11f6cd1819-kube-api-access-j4qn6\") pod \"keystone-cron-29526301-gkh5l\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.474452 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:00 crc kubenswrapper[4492]: I0220 09:01:00.946312 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29526301-gkh5l"] Feb 20 09:01:01 crc kubenswrapper[4492]: I0220 09:01:01.697723 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gkh5l" event={"ID":"c66090c8-b2eb-45ae-9504-2d11f6cd1819","Type":"ContainerStarted","Data":"74802137337a15d129cbb379e0e9c85f29e081e0942144561c4578e5cff2df9d"} Feb 20 09:01:01 crc kubenswrapper[4492]: I0220 09:01:01.700612 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gkh5l" event={"ID":"c66090c8-b2eb-45ae-9504-2d11f6cd1819","Type":"ContainerStarted","Data":"5b3579cda398c72651870ef0263b55f95a752e17832b5f07714add66171424af"} Feb 20 09:01:01 crc kubenswrapper[4492]: I0220 09:01:01.729994 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29526301-gkh5l" podStartSLOduration=1.729976276 podStartE2EDuration="1.729976276s" podCreationTimestamp="2026-02-20 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 09:01:01.715213951 +0000 UTC m=+8418.486502960" watchObservedRunningTime="2026-02-20 09:01:01.729976276 +0000 UTC m=+8418.501265254" Feb 20 09:01:03 crc kubenswrapper[4492]: I0220 09:01:03.718085 4492 generic.go:334] "Generic (PLEG): container finished" podID="c66090c8-b2eb-45ae-9504-2d11f6cd1819" containerID="74802137337a15d129cbb379e0e9c85f29e081e0942144561c4578e5cff2df9d" exitCode=0 Feb 20 09:01:03 crc kubenswrapper[4492]: I0220 09:01:03.718183 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gkh5l" event={"ID":"c66090c8-b2eb-45ae-9504-2d11f6cd1819","Type":"ContainerDied","Data":"74802137337a15d129cbb379e0e9c85f29e081e0942144561c4578e5cff2df9d"} Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.086807 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.168700 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-config-data\") pod \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.168978 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4qn6\" (UniqueName: \"kubernetes.io/projected/c66090c8-b2eb-45ae-9504-2d11f6cd1819-kube-api-access-j4qn6\") pod \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.169154 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-fernet-keys\") pod \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.169397 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-combined-ca-bundle\") pod \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\" (UID: \"c66090c8-b2eb-45ae-9504-2d11f6cd1819\") " Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.176787 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c66090c8-b2eb-45ae-9504-2d11f6cd1819" (UID: "c66090c8-b2eb-45ae-9504-2d11f6cd1819"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.183925 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c66090c8-b2eb-45ae-9504-2d11f6cd1819-kube-api-access-j4qn6" (OuterVolumeSpecName: "kube-api-access-j4qn6") pod "c66090c8-b2eb-45ae-9504-2d11f6cd1819" (UID: "c66090c8-b2eb-45ae-9504-2d11f6cd1819"). InnerVolumeSpecName "kube-api-access-j4qn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.203804 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c66090c8-b2eb-45ae-9504-2d11f6cd1819" (UID: "c66090c8-b2eb-45ae-9504-2d11f6cd1819"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.225689 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-config-data" (OuterVolumeSpecName: "config-data") pod "c66090c8-b2eb-45ae-9504-2d11f6cd1819" (UID: "c66090c8-b2eb-45ae-9504-2d11f6cd1819"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.273063 4492 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.273098 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.273115 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4qn6\" (UniqueName: \"kubernetes.io/projected/c66090c8-b2eb-45ae-9504-2d11f6cd1819-kube-api-access-j4qn6\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.273126 4492 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c66090c8-b2eb-45ae-9504-2d11f6cd1819-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.557965 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:01:05 crc kubenswrapper[4492]: E0220 09:01:05.558311 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.739341 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29526301-gkh5l" event={"ID":"c66090c8-b2eb-45ae-9504-2d11f6cd1819","Type":"ContainerDied","Data":"5b3579cda398c72651870ef0263b55f95a752e17832b5f07714add66171424af"} Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.739400 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b3579cda398c72651870ef0263b55f95a752e17832b5f07714add66171424af" Feb 20 09:01:05 crc kubenswrapper[4492]: I0220 09:01:05.739503 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29526301-gkh5l" Feb 20 09:01:19 crc kubenswrapper[4492]: I0220 09:01:19.557272 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:01:19 crc kubenswrapper[4492]: E0220 09:01:19.558187 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:01:31 crc kubenswrapper[4492]: I0220 09:01:31.558087 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:01:31 crc kubenswrapper[4492]: E0220 09:01:31.559248 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:01:45 crc kubenswrapper[4492]: I0220 09:01:45.557932 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:01:45 crc kubenswrapper[4492]: E0220 09:01:45.559128 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:01:59 crc kubenswrapper[4492]: I0220 09:01:59.564822 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:01:59 crc kubenswrapper[4492]: E0220 09:01:59.565545 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:02:12 crc kubenswrapper[4492]: I0220 09:02:12.558533 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:02:12 crc kubenswrapper[4492]: E0220 09:02:12.559255 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:02:23 crc kubenswrapper[4492]: I0220 09:02:23.563800 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:02:23 crc kubenswrapper[4492]: E0220 09:02:23.564661 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:02:34 crc kubenswrapper[4492]: I0220 09:02:34.557061 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:02:34 crc kubenswrapper[4492]: E0220 09:02:34.557724 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:02:46 crc kubenswrapper[4492]: I0220 09:02:46.557143 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:02:46 crc kubenswrapper[4492]: E0220 09:02:46.558717 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:03:01 crc kubenswrapper[4492]: I0220 09:03:01.557126 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:03:01 crc kubenswrapper[4492]: E0220 09:03:01.557971 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.765600 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f5gzd"] Feb 20 09:03:02 crc kubenswrapper[4492]: E0220 09:03:02.767155 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66090c8-b2eb-45ae-9504-2d11f6cd1819" containerName="keystone-cron" Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.767250 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66090c8-b2eb-45ae-9504-2d11f6cd1819" containerName="keystone-cron" Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.767515 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66090c8-b2eb-45ae-9504-2d11f6cd1819" containerName="keystone-cron" Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.772861 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.804534 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5gzd"] Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.948441 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-catalog-content\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.948548 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-utilities\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:02 crc kubenswrapper[4492]: I0220 09:03:02.948709 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwggj\" (UniqueName: \"kubernetes.io/projected/488157e7-c34a-472b-a0c0-8b21b72eed33-kube-api-access-vwggj\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.051050 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-catalog-content\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.051116 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-utilities\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.051218 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwggj\" (UniqueName: \"kubernetes.io/projected/488157e7-c34a-472b-a0c0-8b21b72eed33-kube-api-access-vwggj\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.052946 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-utilities\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.053081 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-catalog-content\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.086393 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwggj\" (UniqueName: \"kubernetes.io/projected/488157e7-c34a-472b-a0c0-8b21b72eed33-kube-api-access-vwggj\") pod \"redhat-marketplace-f5gzd\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.093710 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.748027 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5gzd"] Feb 20 09:03:03 crc kubenswrapper[4492]: I0220 09:03:03.852359 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5gzd" event={"ID":"488157e7-c34a-472b-a0c0-8b21b72eed33","Type":"ContainerStarted","Data":"84235577e1d99596e35ed6afcdffda1bd8850b0973bb64c6e933f78ed3124499"} Feb 20 09:03:04 crc kubenswrapper[4492]: I0220 09:03:04.864044 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5gzd" event={"ID":"488157e7-c34a-472b-a0c0-8b21b72eed33","Type":"ContainerDied","Data":"fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394"} Feb 20 09:03:04 crc kubenswrapper[4492]: I0220 09:03:04.864832 4492 generic.go:334] "Generic (PLEG): container finished" podID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerID="fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394" exitCode=0 Feb 20 09:03:05 crc kubenswrapper[4492]: I0220 09:03:05.877182 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5gzd" event={"ID":"488157e7-c34a-472b-a0c0-8b21b72eed33","Type":"ContainerStarted","Data":"4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64"} Feb 20 09:03:06 crc kubenswrapper[4492]: I0220 09:03:06.892038 4492 generic.go:334] "Generic (PLEG): container finished" podID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerID="4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64" exitCode=0 Feb 20 09:03:06 crc kubenswrapper[4492]: I0220 09:03:06.892396 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5gzd" event={"ID":"488157e7-c34a-472b-a0c0-8b21b72eed33","Type":"ContainerDied","Data":"4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64"} Feb 20 09:03:07 crc kubenswrapper[4492]: I0220 09:03:07.904636 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5gzd" event={"ID":"488157e7-c34a-472b-a0c0-8b21b72eed33","Type":"ContainerStarted","Data":"b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4"} Feb 20 09:03:07 crc kubenswrapper[4492]: I0220 09:03:07.932353 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f5gzd" podStartSLOduration=3.430171981 podStartE2EDuration="5.930983676s" podCreationTimestamp="2026-02-20 09:03:02 +0000 UTC" firstStartedPulling="2026-02-20 09:03:04.867116866 +0000 UTC m=+8541.638405845" lastFinishedPulling="2026-02-20 09:03:07.367928562 +0000 UTC m=+8544.139217540" observedRunningTime="2026-02-20 09:03:07.921358057 +0000 UTC m=+8544.692647034" watchObservedRunningTime="2026-02-20 09:03:07.930983676 +0000 UTC m=+8544.702272654" Feb 20 09:03:13 crc kubenswrapper[4492]: I0220 09:03:13.094905 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:13 crc kubenswrapper[4492]: I0220 09:03:13.095315 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:13 crc kubenswrapper[4492]: I0220 09:03:13.137121 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:13 crc kubenswrapper[4492]: I0220 09:03:13.997413 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:14 crc kubenswrapper[4492]: I0220 09:03:14.059511 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5gzd"] Feb 20 09:03:15 crc kubenswrapper[4492]: I0220 09:03:15.971300 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f5gzd" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="registry-server" containerID="cri-o://b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4" gracePeriod=2 Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.557983 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:03:16 crc kubenswrapper[4492]: E0220 09:03:16.558270 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.609953 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.667998 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-catalog-content\") pod \"488157e7-c34a-472b-a0c0-8b21b72eed33\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.668241 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwggj\" (UniqueName: \"kubernetes.io/projected/488157e7-c34a-472b-a0c0-8b21b72eed33-kube-api-access-vwggj\") pod \"488157e7-c34a-472b-a0c0-8b21b72eed33\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.668343 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-utilities\") pod \"488157e7-c34a-472b-a0c0-8b21b72eed33\" (UID: \"488157e7-c34a-472b-a0c0-8b21b72eed33\") " Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.670313 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-utilities" (OuterVolumeSpecName: "utilities") pod "488157e7-c34a-472b-a0c0-8b21b72eed33" (UID: "488157e7-c34a-472b-a0c0-8b21b72eed33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.681271 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/488157e7-c34a-472b-a0c0-8b21b72eed33-kube-api-access-vwggj" (OuterVolumeSpecName: "kube-api-access-vwggj") pod "488157e7-c34a-472b-a0c0-8b21b72eed33" (UID: "488157e7-c34a-472b-a0c0-8b21b72eed33"). InnerVolumeSpecName "kube-api-access-vwggj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.691512 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "488157e7-c34a-472b-a0c0-8b21b72eed33" (UID: "488157e7-c34a-472b-a0c0-8b21b72eed33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.773515 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwggj\" (UniqueName: \"kubernetes.io/projected/488157e7-c34a-472b-a0c0-8b21b72eed33-kube-api-access-vwggj\") on node \"crc\" DevicePath \"\"" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.773567 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.773581 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/488157e7-c34a-472b-a0c0-8b21b72eed33-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.983307 4492 generic.go:334] "Generic (PLEG): container finished" podID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerID="b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4" exitCode=0 Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.983405 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5gzd" Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.983393 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5gzd" event={"ID":"488157e7-c34a-472b-a0c0-8b21b72eed33","Type":"ContainerDied","Data":"b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4"} Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.984648 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5gzd" event={"ID":"488157e7-c34a-472b-a0c0-8b21b72eed33","Type":"ContainerDied","Data":"84235577e1d99596e35ed6afcdffda1bd8850b0973bb64c6e933f78ed3124499"} Feb 20 09:03:16 crc kubenswrapper[4492]: I0220 09:03:16.984697 4492 scope.go:117] "RemoveContainer" containerID="b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.024453 4492 scope.go:117] "RemoveContainer" containerID="4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.026261 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5gzd"] Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.048983 4492 scope.go:117] "RemoveContainer" containerID="fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.050404 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5gzd"] Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.084173 4492 scope.go:117] "RemoveContainer" containerID="b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4" Feb 20 09:03:17 crc kubenswrapper[4492]: E0220 09:03:17.086401 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4\": container with ID starting with b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4 not found: ID does not exist" containerID="b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.086448 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4"} err="failed to get container status \"b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4\": rpc error: code = NotFound desc = could not find container \"b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4\": container with ID starting with b0e8066f92bfa722c10eccd6d8006e63dd015d9a3dd5b357777592725da6d5b4 not found: ID does not exist" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.086494 4492 scope.go:117] "RemoveContainer" containerID="4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64" Feb 20 09:03:17 crc kubenswrapper[4492]: E0220 09:03:17.086804 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64\": container with ID starting with 4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64 not found: ID does not exist" containerID="4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.086830 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64"} err="failed to get container status \"4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64\": rpc error: code = NotFound desc = could not find container \"4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64\": container with ID starting with 4570de3627272737be720b736844c63f2fb7c07b46c2b6e86ec5786b86c48c64 not found: ID does not exist" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.086851 4492 scope.go:117] "RemoveContainer" containerID="fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394" Feb 20 09:03:17 crc kubenswrapper[4492]: E0220 09:03:17.087101 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394\": container with ID starting with fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394 not found: ID does not exist" containerID="fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.087128 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394"} err="failed to get container status \"fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394\": rpc error: code = NotFound desc = could not find container \"fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394\": container with ID starting with fd68a50038ef466f9f38687773374123b21bcc3e8d50fdc785fd59a97d493394 not found: ID does not exist" Feb 20 09:03:17 crc kubenswrapper[4492]: E0220 09:03:17.143856 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod488157e7_c34a_472b_a0c0_8b21b72eed33.slice/crio-84235577e1d99596e35ed6afcdffda1bd8850b0973bb64c6e933f78ed3124499\": RecentStats: unable to find data in memory cache]" Feb 20 09:03:17 crc kubenswrapper[4492]: I0220 09:03:17.568912 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" path="/var/lib/kubelet/pods/488157e7-c34a-472b-a0c0-8b21b72eed33/volumes" Feb 20 09:03:28 crc kubenswrapper[4492]: I0220 09:03:28.557639 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:03:28 crc kubenswrapper[4492]: E0220 09:03:28.558697 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:03:40 crc kubenswrapper[4492]: I0220 09:03:40.557639 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:03:40 crc kubenswrapper[4492]: E0220 09:03:40.558535 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:03:55 crc kubenswrapper[4492]: I0220 09:03:55.557924 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:03:55 crc kubenswrapper[4492]: E0220 09:03:55.558891 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:04:10 crc kubenswrapper[4492]: I0220 09:04:10.557086 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:04:10 crc kubenswrapper[4492]: E0220 09:04:10.557831 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:04:24 crc kubenswrapper[4492]: I0220 09:04:24.558221 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:04:24 crc kubenswrapper[4492]: E0220 09:04:24.559179 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:04:36 crc kubenswrapper[4492]: I0220 09:04:36.557500 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:04:36 crc kubenswrapper[4492]: E0220 09:04:36.558704 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:04:49 crc kubenswrapper[4492]: I0220 09:04:49.557060 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:04:49 crc kubenswrapper[4492]: E0220 09:04:49.557727 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:05:03 crc kubenswrapper[4492]: I0220 09:05:03.563086 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:05:03 crc kubenswrapper[4492]: E0220 09:05:03.564733 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:05:15 crc kubenswrapper[4492]: I0220 09:05:15.558171 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:05:15 crc kubenswrapper[4492]: E0220 09:05:15.559020 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.026167 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5bdzl"] Feb 20 09:05:16 crc kubenswrapper[4492]: E0220 09:05:16.028597 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="extract-utilities" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.028625 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="extract-utilities" Feb 20 09:05:16 crc kubenswrapper[4492]: E0220 09:05:16.028683 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="registry-server" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.028692 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="registry-server" Feb 20 09:05:16 crc kubenswrapper[4492]: E0220 09:05:16.028706 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="extract-content" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.028712 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="extract-content" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.029549 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="488157e7-c34a-472b-a0c0-8b21b72eed33" containerName="registry-server" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.032502 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.055777 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5bdzl"] Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.116405 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-catalog-content\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.116788 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-utilities\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.116965 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27q6t\" (UniqueName: \"kubernetes.io/projected/59c477e5-1295-4085-ac57-dc6009e4b5ef-kube-api-access-27q6t\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.218997 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-utilities\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.219075 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27q6t\" (UniqueName: \"kubernetes.io/projected/59c477e5-1295-4085-ac57-dc6009e4b5ef-kube-api-access-27q6t\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.219161 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-catalog-content\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.219765 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-utilities\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.219795 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-catalog-content\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.250131 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27q6t\" (UniqueName: \"kubernetes.io/projected/59c477e5-1295-4085-ac57-dc6009e4b5ef-kube-api-access-27q6t\") pod \"redhat-operators-5bdzl\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.357786 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:16 crc kubenswrapper[4492]: I0220 09:05:16.897195 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5bdzl"] Feb 20 09:05:17 crc kubenswrapper[4492]: I0220 09:05:17.129086 4492 generic.go:334] "Generic (PLEG): container finished" podID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerID="2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0" exitCode=0 Feb 20 09:05:17 crc kubenswrapper[4492]: I0220 09:05:17.129259 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bdzl" event={"ID":"59c477e5-1295-4085-ac57-dc6009e4b5ef","Type":"ContainerDied","Data":"2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0"} Feb 20 09:05:17 crc kubenswrapper[4492]: I0220 09:05:17.129433 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bdzl" event={"ID":"59c477e5-1295-4085-ac57-dc6009e4b5ef","Type":"ContainerStarted","Data":"591b33f0bb85b949b9219666a60facb4f70b524c8672f00209164976d0e4976e"} Feb 20 09:05:17 crc kubenswrapper[4492]: I0220 09:05:17.131343 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:05:18 crc kubenswrapper[4492]: I0220 09:05:18.138964 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bdzl" event={"ID":"59c477e5-1295-4085-ac57-dc6009e4b5ef","Type":"ContainerStarted","Data":"edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47"} Feb 20 09:05:21 crc kubenswrapper[4492]: I0220 09:05:21.168135 4492 generic.go:334] "Generic (PLEG): container finished" podID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerID="edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47" exitCode=0 Feb 20 09:05:21 crc kubenswrapper[4492]: I0220 09:05:21.168233 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bdzl" event={"ID":"59c477e5-1295-4085-ac57-dc6009e4b5ef","Type":"ContainerDied","Data":"edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47"} Feb 20 09:05:22 crc kubenswrapper[4492]: I0220 09:05:22.181349 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bdzl" event={"ID":"59c477e5-1295-4085-ac57-dc6009e4b5ef","Type":"ContainerStarted","Data":"3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404"} Feb 20 09:05:22 crc kubenswrapper[4492]: I0220 09:05:22.207560 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5bdzl" podStartSLOduration=2.679709914 podStartE2EDuration="7.207534007s" podCreationTimestamp="2026-02-20 09:05:15 +0000 UTC" firstStartedPulling="2026-02-20 09:05:17.131071554 +0000 UTC m=+8673.902360532" lastFinishedPulling="2026-02-20 09:05:21.658895646 +0000 UTC m=+8678.430184625" observedRunningTime="2026-02-20 09:05:22.201934349 +0000 UTC m=+8678.973223337" watchObservedRunningTime="2026-02-20 09:05:22.207534007 +0000 UTC m=+8678.978822985" Feb 20 09:05:26 crc kubenswrapper[4492]: I0220 09:05:26.358108 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:26 crc kubenswrapper[4492]: I0220 09:05:26.358644 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:27 crc kubenswrapper[4492]: I0220 09:05:27.408548 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5bdzl" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="registry-server" probeResult="failure" output=< Feb 20 09:05:27 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:05:27 crc kubenswrapper[4492]: > Feb 20 09:05:30 crc kubenswrapper[4492]: I0220 09:05:30.557977 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:05:30 crc kubenswrapper[4492]: E0220 09:05:30.559409 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:05:37 crc kubenswrapper[4492]: I0220 09:05:37.410364 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5bdzl" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="registry-server" probeResult="failure" output=< Feb 20 09:05:37 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:05:37 crc kubenswrapper[4492]: > Feb 20 09:05:45 crc kubenswrapper[4492]: I0220 09:05:45.556928 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:05:46 crc kubenswrapper[4492]: I0220 09:05:46.397458 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:46 crc kubenswrapper[4492]: I0220 09:05:46.457973 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:46 crc kubenswrapper[4492]: I0220 09:05:46.463978 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"eb44ac74a44a21605212c447fda5b5da665b4a874bef8a5cb9d78430fea2dae3"} Feb 20 09:05:47 crc kubenswrapper[4492]: I0220 09:05:47.202291 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5bdzl"] Feb 20 09:05:47 crc kubenswrapper[4492]: I0220 09:05:47.473442 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5bdzl" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="registry-server" containerID="cri-o://3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404" gracePeriod=2 Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.227483 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.419037 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-utilities\") pod \"59c477e5-1295-4085-ac57-dc6009e4b5ef\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.419104 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27q6t\" (UniqueName: \"kubernetes.io/projected/59c477e5-1295-4085-ac57-dc6009e4b5ef-kube-api-access-27q6t\") pod \"59c477e5-1295-4085-ac57-dc6009e4b5ef\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.419333 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-catalog-content\") pod \"59c477e5-1295-4085-ac57-dc6009e4b5ef\" (UID: \"59c477e5-1295-4085-ac57-dc6009e4b5ef\") " Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.421039 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-utilities" (OuterVolumeSpecName: "utilities") pod "59c477e5-1295-4085-ac57-dc6009e4b5ef" (UID: "59c477e5-1295-4085-ac57-dc6009e4b5ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.438114 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59c477e5-1295-4085-ac57-dc6009e4b5ef-kube-api-access-27q6t" (OuterVolumeSpecName: "kube-api-access-27q6t") pod "59c477e5-1295-4085-ac57-dc6009e4b5ef" (UID: "59c477e5-1295-4085-ac57-dc6009e4b5ef"). InnerVolumeSpecName "kube-api-access-27q6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.503990 4492 generic.go:334] "Generic (PLEG): container finished" podID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerID="3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404" exitCode=0 Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.504083 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bdzl" event={"ID":"59c477e5-1295-4085-ac57-dc6009e4b5ef","Type":"ContainerDied","Data":"3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404"} Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.504174 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5bdzl" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.504203 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5bdzl" event={"ID":"59c477e5-1295-4085-ac57-dc6009e4b5ef","Type":"ContainerDied","Data":"591b33f0bb85b949b9219666a60facb4f70b524c8672f00209164976d0e4976e"} Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.504260 4492 scope.go:117] "RemoveContainer" containerID="3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.523195 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27q6t\" (UniqueName: \"kubernetes.io/projected/59c477e5-1295-4085-ac57-dc6009e4b5ef-kube-api-access-27q6t\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.523225 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.532800 4492 scope.go:117] "RemoveContainer" containerID="edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.557933 4492 scope.go:117] "RemoveContainer" containerID="2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.575032 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59c477e5-1295-4085-ac57-dc6009e4b5ef" (UID: "59c477e5-1295-4085-ac57-dc6009e4b5ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.610954 4492 scope.go:117] "RemoveContainer" containerID="3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404" Feb 20 09:05:48 crc kubenswrapper[4492]: E0220 09:05:48.614901 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404\": container with ID starting with 3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404 not found: ID does not exist" containerID="3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.615608 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404"} err="failed to get container status \"3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404\": rpc error: code = NotFound desc = could not find container \"3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404\": container with ID starting with 3bb137f8959faefd6c6a71ffc3bcbd49478b217a02e3c39539ca3fc377268404 not found: ID does not exist" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.615677 4492 scope.go:117] "RemoveContainer" containerID="edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47" Feb 20 09:05:48 crc kubenswrapper[4492]: E0220 09:05:48.617296 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47\": container with ID starting with edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47 not found: ID does not exist" containerID="edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.617335 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47"} err="failed to get container status \"edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47\": rpc error: code = NotFound desc = could not find container \"edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47\": container with ID starting with edecfd41c79757e776469b029510f1873b3033dc6c5ab2340895f099ee2eaf47 not found: ID does not exist" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.617370 4492 scope.go:117] "RemoveContainer" containerID="2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0" Feb 20 09:05:48 crc kubenswrapper[4492]: E0220 09:05:48.618028 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0\": container with ID starting with 2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0 not found: ID does not exist" containerID="2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.618107 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0"} err="failed to get container status \"2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0\": rpc error: code = NotFound desc = could not find container \"2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0\": container with ID starting with 2a3f95c173e42b7d09c28039c4adc7962ce9b5c26d2b87fde262121c32157af0 not found: ID does not exist" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.626574 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59c477e5-1295-4085-ac57-dc6009e4b5ef-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.845811 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5bdzl"] Feb 20 09:05:48 crc kubenswrapper[4492]: I0220 09:05:48.853354 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5bdzl"] Feb 20 09:05:49 crc kubenswrapper[4492]: I0220 09:05:49.580140 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" path="/var/lib/kubelet/pods/59c477e5-1295-4085-ac57-dc6009e4b5ef/volumes" Feb 20 09:08:09 crc kubenswrapper[4492]: I0220 09:08:09.310958 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:08:09 crc kubenswrapper[4492]: I0220 09:08:09.312462 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:08:39 crc kubenswrapper[4492]: I0220 09:08:39.311864 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:08:39 crc kubenswrapper[4492]: I0220 09:08:39.312506 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.311502 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.312243 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.312305 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.313266 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb44ac74a44a21605212c447fda5b5da665b4a874bef8a5cb9d78430fea2dae3"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.313338 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://eb44ac74a44a21605212c447fda5b5da665b4a874bef8a5cb9d78430fea2dae3" gracePeriod=600 Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.540586 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="eb44ac74a44a21605212c447fda5b5da665b4a874bef8a5cb9d78430fea2dae3" exitCode=0 Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.540664 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"eb44ac74a44a21605212c447fda5b5da665b4a874bef8a5cb9d78430fea2dae3"} Feb 20 09:09:09 crc kubenswrapper[4492]: I0220 09:09:09.540797 4492 scope.go:117] "RemoveContainer" containerID="08d79d3288afdf5cb72717776b2eefc320ce33a76bd40a4df88ba2e6e979694d" Feb 20 09:09:10 crc kubenswrapper[4492]: I0220 09:09:10.556255 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b"} Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.295229 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4lvvs"] Feb 20 09:09:26 crc kubenswrapper[4492]: E0220 09:09:26.300317 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="registry-server" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.300350 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="registry-server" Feb 20 09:09:26 crc kubenswrapper[4492]: E0220 09:09:26.300398 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="extract-content" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.300405 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="extract-content" Feb 20 09:09:26 crc kubenswrapper[4492]: E0220 09:09:26.300417 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="extract-utilities" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.300426 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="extract-utilities" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.300945 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="59c477e5-1295-4085-ac57-dc6009e4b5ef" containerName="registry-server" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.306069 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.327217 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4lvvs"] Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.430900 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-utilities\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.431010 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-catalog-content\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.431141 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cngnt\" (UniqueName: \"kubernetes.io/projected/67c8b3b7-c271-4a97-8d85-7573645abd02-kube-api-access-cngnt\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.474048 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-prk8l"] Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.476867 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.496172 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prk8l"] Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.533450 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-utilities\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.533854 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-catalog-content\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.534111 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cngnt\" (UniqueName: \"kubernetes.io/projected/67c8b3b7-c271-4a97-8d85-7573645abd02-kube-api-access-cngnt\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.535174 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-utilities\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.535539 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-catalog-content\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.557819 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cngnt\" (UniqueName: \"kubernetes.io/projected/67c8b3b7-c271-4a97-8d85-7573645abd02-kube-api-access-cngnt\") pod \"certified-operators-4lvvs\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.634400 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.636227 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-catalog-content\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.636460 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-utilities\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.636521 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2td74\" (UniqueName: \"kubernetes.io/projected/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-kube-api-access-2td74\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.738576 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-catalog-content\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.738733 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-utilities\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.738769 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2td74\" (UniqueName: \"kubernetes.io/projected/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-kube-api-access-2td74\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.739288 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-catalog-content\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.739427 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-utilities\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.757784 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2td74\" (UniqueName: \"kubernetes.io/projected/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-kube-api-access-2td74\") pod \"community-operators-prk8l\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:26 crc kubenswrapper[4492]: I0220 09:09:26.819370 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.185671 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4lvvs"] Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.299389 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prk8l"] Feb 20 09:09:27 crc kubenswrapper[4492]: W0220 09:09:27.307612 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode744b0db_b5f8_4dda_b2ea_1abe83d5da71.slice/crio-ee5767e968744d9663c8d84c0b569cc670638445f0ad0a2a1b9de04da2a59a92 WatchSource:0}: Error finding container ee5767e968744d9663c8d84c0b569cc670638445f0ad0a2a1b9de04da2a59a92: Status 404 returned error can't find the container with id ee5767e968744d9663c8d84c0b569cc670638445f0ad0a2a1b9de04da2a59a92 Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.720419 4492 generic.go:334] "Generic (PLEG): container finished" podID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerID="29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc" exitCode=0 Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.720526 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prk8l" event={"ID":"e744b0db-b5f8-4dda-b2ea-1abe83d5da71","Type":"ContainerDied","Data":"29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc"} Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.720786 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prk8l" event={"ID":"e744b0db-b5f8-4dda-b2ea-1abe83d5da71","Type":"ContainerStarted","Data":"ee5767e968744d9663c8d84c0b569cc670638445f0ad0a2a1b9de04da2a59a92"} Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.723874 4492 generic.go:334] "Generic (PLEG): container finished" podID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerID="eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385" exitCode=0 Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.723915 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lvvs" event={"ID":"67c8b3b7-c271-4a97-8d85-7573645abd02","Type":"ContainerDied","Data":"eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385"} Feb 20 09:09:27 crc kubenswrapper[4492]: I0220 09:09:27.723942 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lvvs" event={"ID":"67c8b3b7-c271-4a97-8d85-7573645abd02","Type":"ContainerStarted","Data":"21c55ed1d6fda62cc3810b7c39f80a94d6d86f27c54e29c49b52607974fdc1bc"} Feb 20 09:09:28 crc kubenswrapper[4492]: I0220 09:09:28.738778 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lvvs" event={"ID":"67c8b3b7-c271-4a97-8d85-7573645abd02","Type":"ContainerStarted","Data":"26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2"} Feb 20 09:09:28 crc kubenswrapper[4492]: I0220 09:09:28.745156 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prk8l" event={"ID":"e744b0db-b5f8-4dda-b2ea-1abe83d5da71","Type":"ContainerStarted","Data":"9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f"} Feb 20 09:09:30 crc kubenswrapper[4492]: I0220 09:09:30.783900 4492 generic.go:334] "Generic (PLEG): container finished" podID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerID="9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f" exitCode=0 Feb 20 09:09:30 crc kubenswrapper[4492]: I0220 09:09:30.783992 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prk8l" event={"ID":"e744b0db-b5f8-4dda-b2ea-1abe83d5da71","Type":"ContainerDied","Data":"9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f"} Feb 20 09:09:30 crc kubenswrapper[4492]: I0220 09:09:30.787629 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lvvs" event={"ID":"67c8b3b7-c271-4a97-8d85-7573645abd02","Type":"ContainerDied","Data":"26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2"} Feb 20 09:09:30 crc kubenswrapper[4492]: I0220 09:09:30.787604 4492 generic.go:334] "Generic (PLEG): container finished" podID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerID="26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2" exitCode=0 Feb 20 09:09:31 crc kubenswrapper[4492]: I0220 09:09:31.808007 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lvvs" event={"ID":"67c8b3b7-c271-4a97-8d85-7573645abd02","Type":"ContainerStarted","Data":"e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106"} Feb 20 09:09:31 crc kubenswrapper[4492]: I0220 09:09:31.811562 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prk8l" event={"ID":"e744b0db-b5f8-4dda-b2ea-1abe83d5da71","Type":"ContainerStarted","Data":"fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68"} Feb 20 09:09:31 crc kubenswrapper[4492]: I0220 09:09:31.837104 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4lvvs" podStartSLOduration=2.247530789 podStartE2EDuration="5.836618987s" podCreationTimestamp="2026-02-20 09:09:26 +0000 UTC" firstStartedPulling="2026-02-20 09:09:27.725487273 +0000 UTC m=+8924.496776250" lastFinishedPulling="2026-02-20 09:09:31.31457547 +0000 UTC m=+8928.085864448" observedRunningTime="2026-02-20 09:09:31.829934463 +0000 UTC m=+8928.601223431" watchObservedRunningTime="2026-02-20 09:09:31.836618987 +0000 UTC m=+8928.607907954" Feb 20 09:09:31 crc kubenswrapper[4492]: I0220 09:09:31.862053 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-prk8l" podStartSLOduration=2.321684228 podStartE2EDuration="5.862034468s" podCreationTimestamp="2026-02-20 09:09:26 +0000 UTC" firstStartedPulling="2026-02-20 09:09:27.723147552 +0000 UTC m=+8924.494436530" lastFinishedPulling="2026-02-20 09:09:31.263497792 +0000 UTC m=+8928.034786770" observedRunningTime="2026-02-20 09:09:31.851301361 +0000 UTC m=+8928.622590339" watchObservedRunningTime="2026-02-20 09:09:31.862034468 +0000 UTC m=+8928.633323446" Feb 20 09:09:36 crc kubenswrapper[4492]: I0220 09:09:36.635197 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:36 crc kubenswrapper[4492]: I0220 09:09:36.636043 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:36 crc kubenswrapper[4492]: I0220 09:09:36.820419 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:36 crc kubenswrapper[4492]: I0220 09:09:36.820486 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:37 crc kubenswrapper[4492]: I0220 09:09:37.674014 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-4lvvs" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="registry-server" probeResult="failure" output=< Feb 20 09:09:37 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:09:37 crc kubenswrapper[4492]: > Feb 20 09:09:37 crc kubenswrapper[4492]: I0220 09:09:37.859611 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-prk8l" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="registry-server" probeResult="failure" output=< Feb 20 09:09:37 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:09:37 crc kubenswrapper[4492]: > Feb 20 09:09:46 crc kubenswrapper[4492]: I0220 09:09:46.678225 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:46 crc kubenswrapper[4492]: I0220 09:09:46.731314 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:46 crc kubenswrapper[4492]: I0220 09:09:46.860067 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:46 crc kubenswrapper[4492]: I0220 09:09:46.911040 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:46 crc kubenswrapper[4492]: I0220 09:09:46.912805 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4lvvs"] Feb 20 09:09:47 crc kubenswrapper[4492]: I0220 09:09:47.968728 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4lvvs" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="registry-server" containerID="cri-o://e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106" gracePeriod=2 Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.480277 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.689662 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-utilities\") pod \"67c8b3b7-c271-4a97-8d85-7573645abd02\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.689842 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cngnt\" (UniqueName: \"kubernetes.io/projected/67c8b3b7-c271-4a97-8d85-7573645abd02-kube-api-access-cngnt\") pod \"67c8b3b7-c271-4a97-8d85-7573645abd02\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.689947 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-catalog-content\") pod \"67c8b3b7-c271-4a97-8d85-7573645abd02\" (UID: \"67c8b3b7-c271-4a97-8d85-7573645abd02\") " Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.692046 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-utilities" (OuterVolumeSpecName: "utilities") pod "67c8b3b7-c271-4a97-8d85-7573645abd02" (UID: "67c8b3b7-c271-4a97-8d85-7573645abd02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.697759 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c8b3b7-c271-4a97-8d85-7573645abd02-kube-api-access-cngnt" (OuterVolumeSpecName: "kube-api-access-cngnt") pod "67c8b3b7-c271-4a97-8d85-7573645abd02" (UID: "67c8b3b7-c271-4a97-8d85-7573645abd02"). InnerVolumeSpecName "kube-api-access-cngnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.733121 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67c8b3b7-c271-4a97-8d85-7573645abd02" (UID: "67c8b3b7-c271-4a97-8d85-7573645abd02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.792966 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.792999 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c8b3b7-c271-4a97-8d85-7573645abd02-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.793012 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cngnt\" (UniqueName: \"kubernetes.io/projected/67c8b3b7-c271-4a97-8d85-7573645abd02-kube-api-access-cngnt\") on node \"crc\" DevicePath \"\"" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.983169 4492 generic.go:334] "Generic (PLEG): container finished" podID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerID="e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106" exitCode=0 Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.983214 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lvvs" event={"ID":"67c8b3b7-c271-4a97-8d85-7573645abd02","Type":"ContainerDied","Data":"e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106"} Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.983243 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lvvs" event={"ID":"67c8b3b7-c271-4a97-8d85-7573645abd02","Type":"ContainerDied","Data":"21c55ed1d6fda62cc3810b7c39f80a94d6d86f27c54e29c49b52607974fdc1bc"} Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.983261 4492 scope.go:117] "RemoveContainer" containerID="e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106" Feb 20 09:09:48 crc kubenswrapper[4492]: I0220 09:09:48.983405 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lvvs" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.022748 4492 scope.go:117] "RemoveContainer" containerID="26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.024904 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4lvvs"] Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.037245 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4lvvs"] Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.060876 4492 scope.go:117] "RemoveContainer" containerID="eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.092387 4492 scope.go:117] "RemoveContainer" containerID="e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106" Feb 20 09:09:49 crc kubenswrapper[4492]: E0220 09:09:49.093185 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106\": container with ID starting with e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106 not found: ID does not exist" containerID="e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.093224 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106"} err="failed to get container status \"e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106\": rpc error: code = NotFound desc = could not find container \"e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106\": container with ID starting with e8678ad15f426564c471305b72121fcbcc9a09c69d547e90a18daa0da20af106 not found: ID does not exist" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.093248 4492 scope.go:117] "RemoveContainer" containerID="26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2" Feb 20 09:09:49 crc kubenswrapper[4492]: E0220 09:09:49.093498 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2\": container with ID starting with 26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2 not found: ID does not exist" containerID="26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.093558 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2"} err="failed to get container status \"26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2\": rpc error: code = NotFound desc = could not find container \"26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2\": container with ID starting with 26aa9c2bb40de8328916cd9ae8a3b0a1584fc60e9c2fb36990f006fe1714d3c2 not found: ID does not exist" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.093574 4492 scope.go:117] "RemoveContainer" containerID="eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385" Feb 20 09:09:49 crc kubenswrapper[4492]: E0220 09:09:49.093811 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385\": container with ID starting with eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385 not found: ID does not exist" containerID="eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.093850 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385"} err="failed to get container status \"eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385\": rpc error: code = NotFound desc = could not find container \"eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385\": container with ID starting with eb50e3bcb20aba0752da4e3842d110578f04fed2053228872f80fb5cf74cc385 not found: ID does not exist" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.113016 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prk8l"] Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.114132 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-prk8l" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="registry-server" containerID="cri-o://fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68" gracePeriod=2 Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.511805 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.568173 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" path="/var/lib/kubelet/pods/67c8b3b7-c271-4a97-8d85-7573645abd02/volumes" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.609167 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-utilities\") pod \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.609252 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-catalog-content\") pod \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.609328 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2td74\" (UniqueName: \"kubernetes.io/projected/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-kube-api-access-2td74\") pod \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\" (UID: \"e744b0db-b5f8-4dda-b2ea-1abe83d5da71\") " Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.612042 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-utilities" (OuterVolumeSpecName: "utilities") pod "e744b0db-b5f8-4dda-b2ea-1abe83d5da71" (UID: "e744b0db-b5f8-4dda-b2ea-1abe83d5da71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.620576 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-kube-api-access-2td74" (OuterVolumeSpecName: "kube-api-access-2td74") pod "e744b0db-b5f8-4dda-b2ea-1abe83d5da71" (UID: "e744b0db-b5f8-4dda-b2ea-1abe83d5da71"). InnerVolumeSpecName "kube-api-access-2td74". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.656756 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e744b0db-b5f8-4dda-b2ea-1abe83d5da71" (UID: "e744b0db-b5f8-4dda-b2ea-1abe83d5da71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.712576 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.712628 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.712644 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2td74\" (UniqueName: \"kubernetes.io/projected/e744b0db-b5f8-4dda-b2ea-1abe83d5da71-kube-api-access-2td74\") on node \"crc\" DevicePath \"\"" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.998509 4492 generic.go:334] "Generic (PLEG): container finished" podID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerID="fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68" exitCode=0 Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.998571 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prk8l" event={"ID":"e744b0db-b5f8-4dda-b2ea-1abe83d5da71","Type":"ContainerDied","Data":"fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68"} Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.998612 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prk8l" event={"ID":"e744b0db-b5f8-4dda-b2ea-1abe83d5da71","Type":"ContainerDied","Data":"ee5767e968744d9663c8d84c0b569cc670638445f0ad0a2a1b9de04da2a59a92"} Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.998639 4492 scope.go:117] "RemoveContainer" containerID="fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68" Feb 20 09:09:49 crc kubenswrapper[4492]: I0220 09:09:49.998805 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prk8l" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.036725 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prk8l"] Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.037524 4492 scope.go:117] "RemoveContainer" containerID="9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.050188 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-prk8l"] Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.061878 4492 scope.go:117] "RemoveContainer" containerID="29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.078297 4492 scope.go:117] "RemoveContainer" containerID="fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68" Feb 20 09:09:50 crc kubenswrapper[4492]: E0220 09:09:50.078591 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68\": container with ID starting with fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68 not found: ID does not exist" containerID="fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.078627 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68"} err="failed to get container status \"fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68\": rpc error: code = NotFound desc = could not find container \"fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68\": container with ID starting with fd7a99132c9f93c710150260801bb974af6edf2edaf6057c2d41a49133e78d68 not found: ID does not exist" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.078648 4492 scope.go:117] "RemoveContainer" containerID="9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f" Feb 20 09:09:50 crc kubenswrapper[4492]: E0220 09:09:50.080197 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f\": container with ID starting with 9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f not found: ID does not exist" containerID="9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.080238 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f"} err="failed to get container status \"9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f\": rpc error: code = NotFound desc = could not find container \"9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f\": container with ID starting with 9d4f3e2cf1019004829c5c2b6c32c0fb3e6e7b96d3042877058bee473420f27f not found: ID does not exist" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.080263 4492 scope.go:117] "RemoveContainer" containerID="29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc" Feb 20 09:09:50 crc kubenswrapper[4492]: E0220 09:09:50.080659 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc\": container with ID starting with 29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc not found: ID does not exist" containerID="29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc" Feb 20 09:09:50 crc kubenswrapper[4492]: I0220 09:09:50.080693 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc"} err="failed to get container status \"29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc\": rpc error: code = NotFound desc = could not find container \"29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc\": container with ID starting with 29aba2a99077f4c519a620a0fb609de1aa050784fa5ad0e61a3a534eb7af84cc not found: ID does not exist" Feb 20 09:09:51 crc kubenswrapper[4492]: I0220 09:09:51.567237 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" path="/var/lib/kubelet/pods/e744b0db-b5f8-4dda-b2ea-1abe83d5da71/volumes" Feb 20 09:11:09 crc kubenswrapper[4492]: I0220 09:11:09.311684 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:11:09 crc kubenswrapper[4492]: I0220 09:11:09.312342 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:11:39 crc kubenswrapper[4492]: I0220 09:11:39.311821 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:11:39 crc kubenswrapper[4492]: I0220 09:11:39.312670 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:12:09 crc kubenswrapper[4492]: I0220 09:12:09.311501 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:12:09 crc kubenswrapper[4492]: I0220 09:12:09.312277 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:12:09 crc kubenswrapper[4492]: I0220 09:12:09.312354 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 09:12:09 crc kubenswrapper[4492]: I0220 09:12:09.313784 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:12:09 crc kubenswrapper[4492]: I0220 09:12:09.313869 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" gracePeriod=600 Feb 20 09:12:09 crc kubenswrapper[4492]: E0220 09:12:09.457016 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:12:09 crc kubenswrapper[4492]: E0220 09:12:09.544031 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8755354_8581_4b5d_9b7e_6b7c53707f8c.slice/crio-bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8755354_8581_4b5d_9b7e_6b7c53707f8c.slice/crio-conmon-bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b.scope\": RecentStats: unable to find data in memory cache]" Feb 20 09:12:10 crc kubenswrapper[4492]: I0220 09:12:10.387713 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" exitCode=0 Feb 20 09:12:10 crc kubenswrapper[4492]: I0220 09:12:10.387786 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b"} Feb 20 09:12:10 crc kubenswrapper[4492]: I0220 09:12:10.388181 4492 scope.go:117] "RemoveContainer" containerID="eb44ac74a44a21605212c447fda5b5da665b4a874bef8a5cb9d78430fea2dae3" Feb 20 09:12:10 crc kubenswrapper[4492]: I0220 09:12:10.389605 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:12:10 crc kubenswrapper[4492]: E0220 09:12:10.390124 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:12:23 crc kubenswrapper[4492]: I0220 09:12:23.564529 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:12:23 crc kubenswrapper[4492]: E0220 09:12:23.565286 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:12:35 crc kubenswrapper[4492]: I0220 09:12:35.557237 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:12:35 crc kubenswrapper[4492]: E0220 09:12:35.563069 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:12:50 crc kubenswrapper[4492]: I0220 09:12:50.557568 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:12:50 crc kubenswrapper[4492]: E0220 09:12:50.558805 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:13:04 crc kubenswrapper[4492]: I0220 09:13:04.556739 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:13:04 crc kubenswrapper[4492]: E0220 09:13:04.557797 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:13:19 crc kubenswrapper[4492]: I0220 09:13:19.557738 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:13:19 crc kubenswrapper[4492]: E0220 09:13:19.558939 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.759113 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhfj"] Feb 20 09:13:31 crc kubenswrapper[4492]: E0220 09:13:31.761413 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="extract-content" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.761438 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="extract-content" Feb 20 09:13:31 crc kubenswrapper[4492]: E0220 09:13:31.761450 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="extract-utilities" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.761458 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="extract-utilities" Feb 20 09:13:31 crc kubenswrapper[4492]: E0220 09:13:31.761485 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="registry-server" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.761491 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="registry-server" Feb 20 09:13:31 crc kubenswrapper[4492]: E0220 09:13:31.761505 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="extract-utilities" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.761511 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="extract-utilities" Feb 20 09:13:31 crc kubenswrapper[4492]: E0220 09:13:31.761525 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="registry-server" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.761531 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="registry-server" Feb 20 09:13:31 crc kubenswrapper[4492]: E0220 09:13:31.761554 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="extract-content" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.761560 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="extract-content" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.762226 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="e744b0db-b5f8-4dda-b2ea-1abe83d5da71" containerName="registry-server" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.762244 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c8b3b7-c271-4a97-8d85-7573645abd02" containerName="registry-server" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.764047 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.773935 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhfj"] Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.814600 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjcv2\" (UniqueName: \"kubernetes.io/projected/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-kube-api-access-jjcv2\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.815124 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-utilities\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.815526 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-catalog-content\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.917171 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-catalog-content\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.917361 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjcv2\" (UniqueName: \"kubernetes.io/projected/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-kube-api-access-jjcv2\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.917487 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-utilities\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.917874 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-catalog-content\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.918052 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-utilities\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:31 crc kubenswrapper[4492]: I0220 09:13:31.944097 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjcv2\" (UniqueName: \"kubernetes.io/projected/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-kube-api-access-jjcv2\") pod \"redhat-marketplace-pfhfj\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:32 crc kubenswrapper[4492]: I0220 09:13:32.094572 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:32 crc kubenswrapper[4492]: I0220 09:13:32.617032 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhfj"] Feb 20 09:13:33 crc kubenswrapper[4492]: I0220 09:13:33.212601 4492 generic.go:334] "Generic (PLEG): container finished" podID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerID="c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0" exitCode=0 Feb 20 09:13:33 crc kubenswrapper[4492]: I0220 09:13:33.212706 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhfj" event={"ID":"1e1dd0bb-f857-4e1d-8646-f98503a70ca8","Type":"ContainerDied","Data":"c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0"} Feb 20 09:13:33 crc kubenswrapper[4492]: I0220 09:13:33.213050 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhfj" event={"ID":"1e1dd0bb-f857-4e1d-8646-f98503a70ca8","Type":"ContainerStarted","Data":"7c646548ec6aea3c1caf92fc3151ebf232de8c918237c9bd3a4c03571988044c"} Feb 20 09:13:33 crc kubenswrapper[4492]: I0220 09:13:33.222256 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:13:33 crc kubenswrapper[4492]: I0220 09:13:33.567266 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:13:33 crc kubenswrapper[4492]: E0220 09:13:33.567703 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:13:34 crc kubenswrapper[4492]: I0220 09:13:34.224769 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhfj" event={"ID":"1e1dd0bb-f857-4e1d-8646-f98503a70ca8","Type":"ContainerStarted","Data":"dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6"} Feb 20 09:13:35 crc kubenswrapper[4492]: I0220 09:13:35.238326 4492 generic.go:334] "Generic (PLEG): container finished" podID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerID="dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6" exitCode=0 Feb 20 09:13:35 crc kubenswrapper[4492]: I0220 09:13:35.238457 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhfj" event={"ID":"1e1dd0bb-f857-4e1d-8646-f98503a70ca8","Type":"ContainerDied","Data":"dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6"} Feb 20 09:13:37 crc kubenswrapper[4492]: I0220 09:13:37.263424 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhfj" event={"ID":"1e1dd0bb-f857-4e1d-8646-f98503a70ca8","Type":"ContainerStarted","Data":"64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422"} Feb 20 09:13:37 crc kubenswrapper[4492]: I0220 09:13:37.286179 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pfhfj" podStartSLOduration=3.769020014 podStartE2EDuration="6.286157442s" podCreationTimestamp="2026-02-20 09:13:31 +0000 UTC" firstStartedPulling="2026-02-20 09:13:33.219525846 +0000 UTC m=+9169.990814824" lastFinishedPulling="2026-02-20 09:13:35.736663274 +0000 UTC m=+9172.507952252" observedRunningTime="2026-02-20 09:13:37.280427107 +0000 UTC m=+9174.051716086" watchObservedRunningTime="2026-02-20 09:13:37.286157442 +0000 UTC m=+9174.057446421" Feb 20 09:13:42 crc kubenswrapper[4492]: I0220 09:13:42.095696 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:42 crc kubenswrapper[4492]: I0220 09:13:42.096576 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:42 crc kubenswrapper[4492]: I0220 09:13:42.140893 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:42 crc kubenswrapper[4492]: I0220 09:13:42.356007 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:42 crc kubenswrapper[4492]: I0220 09:13:42.414869 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhfj"] Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.329622 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pfhfj" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="registry-server" containerID="cri-o://64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422" gracePeriod=2 Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.557834 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:13:44 crc kubenswrapper[4492]: E0220 09:13:44.558145 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.840696 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.910959 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjcv2\" (UniqueName: \"kubernetes.io/projected/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-kube-api-access-jjcv2\") pod \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.911002 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-catalog-content\") pod \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.911192 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-utilities\") pod \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\" (UID: \"1e1dd0bb-f857-4e1d-8646-f98503a70ca8\") " Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.912314 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-utilities" (OuterVolumeSpecName: "utilities") pod "1e1dd0bb-f857-4e1d-8646-f98503a70ca8" (UID: "1e1dd0bb-f857-4e1d-8646-f98503a70ca8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.924830 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-kube-api-access-jjcv2" (OuterVolumeSpecName: "kube-api-access-jjcv2") pod "1e1dd0bb-f857-4e1d-8646-f98503a70ca8" (UID: "1e1dd0bb-f857-4e1d-8646-f98503a70ca8"). InnerVolumeSpecName "kube-api-access-jjcv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:13:44 crc kubenswrapper[4492]: I0220 09:13:44.934609 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e1dd0bb-f857-4e1d-8646-f98503a70ca8" (UID: "1e1dd0bb-f857-4e1d-8646-f98503a70ca8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.013937 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.013978 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjcv2\" (UniqueName: \"kubernetes.io/projected/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-kube-api-access-jjcv2\") on node \"crc\" DevicePath \"\"" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.013990 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e1dd0bb-f857-4e1d-8646-f98503a70ca8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.343168 4492 generic.go:334] "Generic (PLEG): container finished" podID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerID="64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422" exitCode=0 Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.343224 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhfj" event={"ID":"1e1dd0bb-f857-4e1d-8646-f98503a70ca8","Type":"ContainerDied","Data":"64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422"} Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.343260 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfhfj" event={"ID":"1e1dd0bb-f857-4e1d-8646-f98503a70ca8","Type":"ContainerDied","Data":"7c646548ec6aea3c1caf92fc3151ebf232de8c918237c9bd3a4c03571988044c"} Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.343268 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfhfj" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.343283 4492 scope.go:117] "RemoveContainer" containerID="64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.379774 4492 scope.go:117] "RemoveContainer" containerID="dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.382875 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhfj"] Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.397817 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfhfj"] Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.405598 4492 scope.go:117] "RemoveContainer" containerID="c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.449688 4492 scope.go:117] "RemoveContainer" containerID="64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422" Feb 20 09:13:45 crc kubenswrapper[4492]: E0220 09:13:45.450231 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422\": container with ID starting with 64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422 not found: ID does not exist" containerID="64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.450269 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422"} err="failed to get container status \"64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422\": rpc error: code = NotFound desc = could not find container \"64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422\": container with ID starting with 64e2aad1e0c6dc3963208992806e2110108dca7df703c88749da66e7e67d0422 not found: ID does not exist" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.450298 4492 scope.go:117] "RemoveContainer" containerID="dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6" Feb 20 09:13:45 crc kubenswrapper[4492]: E0220 09:13:45.451038 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6\": container with ID starting with dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6 not found: ID does not exist" containerID="dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.451062 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6"} err="failed to get container status \"dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6\": rpc error: code = NotFound desc = could not find container \"dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6\": container with ID starting with dfe111ad51fcc626f5de2315dab7710a966d6f86de7ba503540b9fe434b418a6 not found: ID does not exist" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.451087 4492 scope.go:117] "RemoveContainer" containerID="c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0" Feb 20 09:13:45 crc kubenswrapper[4492]: E0220 09:13:45.454180 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0\": container with ID starting with c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0 not found: ID does not exist" containerID="c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.454204 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0"} err="failed to get container status \"c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0\": rpc error: code = NotFound desc = could not find container \"c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0\": container with ID starting with c6b4ce941143720e77640d4e400f12d217a6645f2f3fab8e474cc3e02899b5f0 not found: ID does not exist" Feb 20 09:13:45 crc kubenswrapper[4492]: I0220 09:13:45.569043 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" path="/var/lib/kubelet/pods/1e1dd0bb-f857-4e1d-8646-f98503a70ca8/volumes" Feb 20 09:13:58 crc kubenswrapper[4492]: I0220 09:13:58.557594 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:13:58 crc kubenswrapper[4492]: E0220 09:13:58.558611 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:14:10 crc kubenswrapper[4492]: I0220 09:14:10.557409 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:14:10 crc kubenswrapper[4492]: E0220 09:14:10.559290 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:14:24 crc kubenswrapper[4492]: I0220 09:14:24.557072 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:14:24 crc kubenswrapper[4492]: E0220 09:14:24.558147 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:14:38 crc kubenswrapper[4492]: I0220 09:14:38.557129 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:14:38 crc kubenswrapper[4492]: E0220 09:14:38.558815 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:14:49 crc kubenswrapper[4492]: I0220 09:14:49.557708 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:14:49 crc kubenswrapper[4492]: E0220 09:14:49.558655 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.220248 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd"] Feb 20 09:15:00 crc kubenswrapper[4492]: E0220 09:15:00.221284 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="extract-content" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.221301 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="extract-content" Feb 20 09:15:00 crc kubenswrapper[4492]: E0220 09:15:00.221331 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.221336 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4492]: E0220 09:15:00.221348 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="extract-utilities" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.221354 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="extract-utilities" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.221624 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e1dd0bb-f857-4e1d-8646-f98503a70ca8" containerName="registry-server" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.222360 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.232665 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.301283 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.310259 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd"] Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.396609 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txpfd\" (UniqueName: \"kubernetes.io/projected/83f12ab0-70e8-44b7-a470-9d5ca615d344-kube-api-access-txpfd\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.396787 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f12ab0-70e8-44b7-a470-9d5ca615d344-config-volume\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.397015 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f12ab0-70e8-44b7-a470-9d5ca615d344-secret-volume\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.498821 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txpfd\" (UniqueName: \"kubernetes.io/projected/83f12ab0-70e8-44b7-a470-9d5ca615d344-kube-api-access-txpfd\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.498871 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f12ab0-70e8-44b7-a470-9d5ca615d344-config-volume\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.498919 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f12ab0-70e8-44b7-a470-9d5ca615d344-secret-volume\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.500186 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f12ab0-70e8-44b7-a470-9d5ca615d344-config-volume\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.506188 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f12ab0-70e8-44b7-a470-9d5ca615d344-secret-volume\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.516593 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txpfd\" (UniqueName: \"kubernetes.io/projected/83f12ab0-70e8-44b7-a470-9d5ca615d344-kube-api-access-txpfd\") pod \"collect-profiles-29526315-mhdsd\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:00 crc kubenswrapper[4492]: I0220 09:15:00.545777 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:01 crc kubenswrapper[4492]: I0220 09:15:01.127981 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd"] Feb 20 09:15:02 crc kubenswrapper[4492]: I0220 09:15:02.080960 4492 generic.go:334] "Generic (PLEG): container finished" podID="83f12ab0-70e8-44b7-a470-9d5ca615d344" containerID="4c1df81e4ed67e83956b9c07ca6fdf3b4323e74a88675ff37470f4b3a5e74b44" exitCode=0 Feb 20 09:15:02 crc kubenswrapper[4492]: I0220 09:15:02.081066 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" event={"ID":"83f12ab0-70e8-44b7-a470-9d5ca615d344","Type":"ContainerDied","Data":"4c1df81e4ed67e83956b9c07ca6fdf3b4323e74a88675ff37470f4b3a5e74b44"} Feb 20 09:15:02 crc kubenswrapper[4492]: I0220 09:15:02.081668 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" event={"ID":"83f12ab0-70e8-44b7-a470-9d5ca615d344","Type":"ContainerStarted","Data":"1235d541892d5e354d0c8a2c3ee699912a9debe8596fe49a1ab433005af7439d"} Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.391930 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.476847 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txpfd\" (UniqueName: \"kubernetes.io/projected/83f12ab0-70e8-44b7-a470-9d5ca615d344-kube-api-access-txpfd\") pod \"83f12ab0-70e8-44b7-a470-9d5ca615d344\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.477254 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f12ab0-70e8-44b7-a470-9d5ca615d344-config-volume\") pod \"83f12ab0-70e8-44b7-a470-9d5ca615d344\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.477451 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f12ab0-70e8-44b7-a470-9d5ca615d344-secret-volume\") pod \"83f12ab0-70e8-44b7-a470-9d5ca615d344\" (UID: \"83f12ab0-70e8-44b7-a470-9d5ca615d344\") " Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.477946 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83f12ab0-70e8-44b7-a470-9d5ca615d344-config-volume" (OuterVolumeSpecName: "config-volume") pod "83f12ab0-70e8-44b7-a470-9d5ca615d344" (UID: "83f12ab0-70e8-44b7-a470-9d5ca615d344"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.478738 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83f12ab0-70e8-44b7-a470-9d5ca615d344-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.486376 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f12ab0-70e8-44b7-a470-9d5ca615d344-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "83f12ab0-70e8-44b7-a470-9d5ca615d344" (UID: "83f12ab0-70e8-44b7-a470-9d5ca615d344"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.489710 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83f12ab0-70e8-44b7-a470-9d5ca615d344-kube-api-access-txpfd" (OuterVolumeSpecName: "kube-api-access-txpfd") pod "83f12ab0-70e8-44b7-a470-9d5ca615d344" (UID: "83f12ab0-70e8-44b7-a470-9d5ca615d344"). InnerVolumeSpecName "kube-api-access-txpfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.579939 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83f12ab0-70e8-44b7-a470-9d5ca615d344-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:03 crc kubenswrapper[4492]: I0220 09:15:03.580082 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txpfd\" (UniqueName: \"kubernetes.io/projected/83f12ab0-70e8-44b7-a470-9d5ca615d344-kube-api-access-txpfd\") on node \"crc\" DevicePath \"\"" Feb 20 09:15:04 crc kubenswrapper[4492]: I0220 09:15:04.099539 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" event={"ID":"83f12ab0-70e8-44b7-a470-9d5ca615d344","Type":"ContainerDied","Data":"1235d541892d5e354d0c8a2c3ee699912a9debe8596fe49a1ab433005af7439d"} Feb 20 09:15:04 crc kubenswrapper[4492]: I0220 09:15:04.099971 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1235d541892d5e354d0c8a2c3ee699912a9debe8596fe49a1ab433005af7439d" Feb 20 09:15:04 crc kubenswrapper[4492]: I0220 09:15:04.099752 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526315-mhdsd" Feb 20 09:15:04 crc kubenswrapper[4492]: I0220 09:15:04.488763 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh"] Feb 20 09:15:04 crc kubenswrapper[4492]: I0220 09:15:04.497878 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526270-clhlh"] Feb 20 09:15:04 crc kubenswrapper[4492]: I0220 09:15:04.557939 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:15:04 crc kubenswrapper[4492]: E0220 09:15:04.558687 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:15:05 crc kubenswrapper[4492]: I0220 09:15:05.570858 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae3b0337-8c14-4f19-9c88-6da340069e0d" path="/var/lib/kubelet/pods/ae3b0337-8c14-4f19-9c88-6da340069e0d/volumes" Feb 20 09:15:11 crc kubenswrapper[4492]: I0220 09:15:11.147181 4492 scope.go:117] "RemoveContainer" containerID="ff8c07c9b1493eb4cb24e4a14e42f263f201df801d495b6fcc5c57d09a9caaac" Feb 20 09:15:17 crc kubenswrapper[4492]: I0220 09:15:17.567292 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:15:17 crc kubenswrapper[4492]: E0220 09:15:17.576732 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:15:32 crc kubenswrapper[4492]: I0220 09:15:32.556806 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:15:32 crc kubenswrapper[4492]: E0220 09:15:32.557785 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:15:47 crc kubenswrapper[4492]: I0220 09:15:47.557309 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:15:47 crc kubenswrapper[4492]: E0220 09:15:47.558355 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.597056 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t2cjj"] Feb 20 09:15:50 crc kubenswrapper[4492]: E0220 09:15:50.600126 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f12ab0-70e8-44b7-a470-9d5ca615d344" containerName="collect-profiles" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.600169 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f12ab0-70e8-44b7-a470-9d5ca615d344" containerName="collect-profiles" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.601198 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f12ab0-70e8-44b7-a470-9d5ca615d344" containerName="collect-profiles" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.605325 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.628423 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2cjj"] Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.698457 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.699323 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-utilities\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.699457 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbchz\" (UniqueName: \"kubernetes.io/projected/f996a4ab-3306-4226-9e29-6c14e3b22867-kube-api-access-qbchz\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.801684 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-utilities\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.801994 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbchz\" (UniqueName: \"kubernetes.io/projected/f996a4ab-3306-4226-9e29-6c14e3b22867-kube-api-access-qbchz\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.802164 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.802964 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-utilities\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.803578 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.830198 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbchz\" (UniqueName: \"kubernetes.io/projected/f996a4ab-3306-4226-9e29-6c14e3b22867-kube-api-access-qbchz\") pod \"redhat-operators-t2cjj\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:50 crc kubenswrapper[4492]: I0220 09:15:50.931094 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:15:51 crc kubenswrapper[4492]: I0220 09:15:51.713425 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2cjj"] Feb 20 09:15:52 crc kubenswrapper[4492]: I0220 09:15:52.557519 4492 generic.go:334] "Generic (PLEG): container finished" podID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerID="840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8" exitCode=0 Feb 20 09:15:52 crc kubenswrapper[4492]: I0220 09:15:52.557615 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2cjj" event={"ID":"f996a4ab-3306-4226-9e29-6c14e3b22867","Type":"ContainerDied","Data":"840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8"} Feb 20 09:15:52 crc kubenswrapper[4492]: I0220 09:15:52.557818 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2cjj" event={"ID":"f996a4ab-3306-4226-9e29-6c14e3b22867","Type":"ContainerStarted","Data":"d9ef96e6343b5cbe7276e8749727c4baf85b761ba4b78ac174885fe32f898671"} Feb 20 09:15:53 crc kubenswrapper[4492]: I0220 09:15:53.573413 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2cjj" event={"ID":"f996a4ab-3306-4226-9e29-6c14e3b22867","Type":"ContainerStarted","Data":"77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c"} Feb 20 09:15:56 crc kubenswrapper[4492]: I0220 09:15:56.601224 4492 generic.go:334] "Generic (PLEG): container finished" podID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerID="77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c" exitCode=0 Feb 20 09:15:56 crc kubenswrapper[4492]: I0220 09:15:56.601334 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2cjj" event={"ID":"f996a4ab-3306-4226-9e29-6c14e3b22867","Type":"ContainerDied","Data":"77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c"} Feb 20 09:15:58 crc kubenswrapper[4492]: I0220 09:15:58.620649 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2cjj" event={"ID":"f996a4ab-3306-4226-9e29-6c14e3b22867","Type":"ContainerStarted","Data":"2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34"} Feb 20 09:15:58 crc kubenswrapper[4492]: I0220 09:15:58.647168 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t2cjj" podStartSLOduration=4.130816708 podStartE2EDuration="8.646123944s" podCreationTimestamp="2026-02-20 09:15:50 +0000 UTC" firstStartedPulling="2026-02-20 09:15:52.559135915 +0000 UTC m=+9309.330424893" lastFinishedPulling="2026-02-20 09:15:57.074443151 +0000 UTC m=+9313.845732129" observedRunningTime="2026-02-20 09:15:58.645932733 +0000 UTC m=+9315.417221712" watchObservedRunningTime="2026-02-20 09:15:58.646123944 +0000 UTC m=+9315.417412921" Feb 20 09:16:00 crc kubenswrapper[4492]: I0220 09:16:00.558585 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:16:00 crc kubenswrapper[4492]: E0220 09:16:00.559118 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:16:00 crc kubenswrapper[4492]: I0220 09:16:00.932088 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:16:00 crc kubenswrapper[4492]: I0220 09:16:00.932179 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:16:01 crc kubenswrapper[4492]: I0220 09:16:01.979460 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t2cjj" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="registry-server" probeResult="failure" output=< Feb 20 09:16:01 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:16:01 crc kubenswrapper[4492]: > Feb 20 09:16:11 crc kubenswrapper[4492]: I0220 09:16:11.972512 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t2cjj" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="registry-server" probeResult="failure" output=< Feb 20 09:16:11 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:16:11 crc kubenswrapper[4492]: > Feb 20 09:16:12 crc kubenswrapper[4492]: I0220 09:16:12.557252 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:16:12 crc kubenswrapper[4492]: E0220 09:16:12.557585 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:16:21 crc kubenswrapper[4492]: I0220 09:16:21.344151 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:16:21 crc kubenswrapper[4492]: I0220 09:16:21.391620 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:16:21 crc kubenswrapper[4492]: I0220 09:16:21.812286 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t2cjj"] Feb 20 09:16:22 crc kubenswrapper[4492]: I0220 09:16:22.870772 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t2cjj" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="registry-server" containerID="cri-o://2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34" gracePeriod=2 Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.568676 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:16:23 crc kubenswrapper[4492]: E0220 09:16:23.569802 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.618247 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.779500 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-utilities\") pod \"f996a4ab-3306-4226-9e29-6c14e3b22867\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.779998 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbchz\" (UniqueName: \"kubernetes.io/projected/f996a4ab-3306-4226-9e29-6c14e3b22867-kube-api-access-qbchz\") pod \"f996a4ab-3306-4226-9e29-6c14e3b22867\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.780137 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content\") pod \"f996a4ab-3306-4226-9e29-6c14e3b22867\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.781151 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-utilities" (OuterVolumeSpecName: "utilities") pod "f996a4ab-3306-4226-9e29-6c14e3b22867" (UID: "f996a4ab-3306-4226-9e29-6c14e3b22867"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.798821 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f996a4ab-3306-4226-9e29-6c14e3b22867-kube-api-access-qbchz" (OuterVolumeSpecName: "kube-api-access-qbchz") pod "f996a4ab-3306-4226-9e29-6c14e3b22867" (UID: "f996a4ab-3306-4226-9e29-6c14e3b22867"). InnerVolumeSpecName "kube-api-access-qbchz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.881241 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f996a4ab-3306-4226-9e29-6c14e3b22867" (UID: "f996a4ab-3306-4226-9e29-6c14e3b22867"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.882794 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content\") pod \"f996a4ab-3306-4226-9e29-6c14e3b22867\" (UID: \"f996a4ab-3306-4226-9e29-6c14e3b22867\") " Feb 20 09:16:23 crc kubenswrapper[4492]: W0220 09:16:23.884297 4492 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/f996a4ab-3306-4226-9e29-6c14e3b22867/volumes/kubernetes.io~empty-dir/catalog-content Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.884923 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f996a4ab-3306-4226-9e29-6c14e3b22867" (UID: "f996a4ab-3306-4226-9e29-6c14e3b22867"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886303 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886344 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbchz\" (UniqueName: \"kubernetes.io/projected/f996a4ab-3306-4226-9e29-6c14e3b22867-kube-api-access-qbchz\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886359 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f996a4ab-3306-4226-9e29-6c14e3b22867-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886379 4492 generic.go:334] "Generic (PLEG): container finished" podID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerID="2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34" exitCode=0 Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886428 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2cjj" event={"ID":"f996a4ab-3306-4226-9e29-6c14e3b22867","Type":"ContainerDied","Data":"2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34"} Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886569 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2cjj" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886604 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2cjj" event={"ID":"f996a4ab-3306-4226-9e29-6c14e3b22867","Type":"ContainerDied","Data":"d9ef96e6343b5cbe7276e8749727c4baf85b761ba4b78ac174885fe32f898671"} Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.886634 4492 scope.go:117] "RemoveContainer" containerID="2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.926244 4492 scope.go:117] "RemoveContainer" containerID="77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.931928 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t2cjj"] Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.950013 4492 scope.go:117] "RemoveContainer" containerID="840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.974194 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t2cjj"] Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.992693 4492 scope.go:117] "RemoveContainer" containerID="2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34" Feb 20 09:16:23 crc kubenswrapper[4492]: E0220 09:16:23.994879 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34\": container with ID starting with 2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34 not found: ID does not exist" containerID="2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.994986 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34"} err="failed to get container status \"2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34\": rpc error: code = NotFound desc = could not find container \"2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34\": container with ID starting with 2c70b92430ec28154a772bf05abfa8cb9788067d51ae39055eabd984376e9d34 not found: ID does not exist" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.995020 4492 scope.go:117] "RemoveContainer" containerID="77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c" Feb 20 09:16:23 crc kubenswrapper[4492]: E0220 09:16:23.995364 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c\": container with ID starting with 77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c not found: ID does not exist" containerID="77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.995401 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c"} err="failed to get container status \"77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c\": rpc error: code = NotFound desc = could not find container \"77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c\": container with ID starting with 77fcf802125210b54b3301985018895f24899be2e8a97ba617c3baecec6e535c not found: ID does not exist" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.995425 4492 scope.go:117] "RemoveContainer" containerID="840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8" Feb 20 09:16:23 crc kubenswrapper[4492]: E0220 09:16:23.995696 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8\": container with ID starting with 840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8 not found: ID does not exist" containerID="840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8" Feb 20 09:16:23 crc kubenswrapper[4492]: I0220 09:16:23.995732 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8"} err="failed to get container status \"840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8\": rpc error: code = NotFound desc = could not find container \"840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8\": container with ID starting with 840ea952c653f5ef06b6b8619d5c0e75f94067e1f989f20e0ad308a2550894e8 not found: ID does not exist" Feb 20 09:16:25 crc kubenswrapper[4492]: I0220 09:16:25.575878 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" path="/var/lib/kubelet/pods/f996a4ab-3306-4226-9e29-6c14e3b22867/volumes" Feb 20 09:16:38 crc kubenswrapper[4492]: I0220 09:16:38.557633 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:16:38 crc kubenswrapper[4492]: E0220 09:16:38.558595 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:16:49 crc kubenswrapper[4492]: I0220 09:16:49.557147 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:16:49 crc kubenswrapper[4492]: E0220 09:16:49.558088 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:17:02 crc kubenswrapper[4492]: I0220 09:17:02.557988 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:17:02 crc kubenswrapper[4492]: E0220 09:17:02.562385 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:17:16 crc kubenswrapper[4492]: I0220 09:17:16.557185 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:17:17 crc kubenswrapper[4492]: I0220 09:17:17.397414 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"6ff6c52a7e322eb6e1d90878291c8341550e11afc0ae66488e8cec367610b76a"} Feb 20 09:19:35 crc kubenswrapper[4492]: I0220 09:19:35.960240 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jtr54"] Feb 20 09:19:35 crc kubenswrapper[4492]: E0220 09:19:35.971754 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="extract-utilities" Feb 20 09:19:35 crc kubenswrapper[4492]: I0220 09:19:35.971799 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="extract-utilities" Feb 20 09:19:35 crc kubenswrapper[4492]: E0220 09:19:35.971845 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="extract-content" Feb 20 09:19:35 crc kubenswrapper[4492]: I0220 09:19:35.971853 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="extract-content" Feb 20 09:19:35 crc kubenswrapper[4492]: E0220 09:19:35.971900 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="registry-server" Feb 20 09:19:35 crc kubenswrapper[4492]: I0220 09:19:35.971907 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="registry-server" Feb 20 09:19:35 crc kubenswrapper[4492]: I0220 09:19:35.975148 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f996a4ab-3306-4226-9e29-6c14e3b22867" containerName="registry-server" Feb 20 09:19:35 crc kubenswrapper[4492]: I0220 09:19:35.978789 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:35 crc kubenswrapper[4492]: I0220 09:19:35.984033 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtr54"] Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.170685 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-utilities\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.170757 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srpnm\" (UniqueName: \"kubernetes.io/projected/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-kube-api-access-srpnm\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.170799 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-catalog-content\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.272686 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-utilities\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.272755 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srpnm\" (UniqueName: \"kubernetes.io/projected/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-kube-api-access-srpnm\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.272785 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-catalog-content\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.273356 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-catalog-content\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.273870 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-utilities\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.716435 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srpnm\" (UniqueName: \"kubernetes.io/projected/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-kube-api-access-srpnm\") pod \"certified-operators-jtr54\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:36 crc kubenswrapper[4492]: I0220 09:19:36.914265 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:37 crc kubenswrapper[4492]: I0220 09:19:37.591678 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jtr54"] Feb 20 09:19:37 crc kubenswrapper[4492]: I0220 09:19:37.722506 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtr54" event={"ID":"f4c8b933-df50-49a0-8c5f-f2d905a7be1c","Type":"ContainerStarted","Data":"e67b7b278a96ee55d66871920be1a47602848574c5597686fc77f2632ec70241"} Feb 20 09:19:38 crc kubenswrapper[4492]: I0220 09:19:38.730167 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtr54" event={"ID":"f4c8b933-df50-49a0-8c5f-f2d905a7be1c","Type":"ContainerDied","Data":"014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666"} Feb 20 09:19:38 crc kubenswrapper[4492]: I0220 09:19:38.731137 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerID="014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666" exitCode=0 Feb 20 09:19:38 crc kubenswrapper[4492]: I0220 09:19:38.734774 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:19:39 crc kubenswrapper[4492]: I0220 09:19:39.311514 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:19:39 crc kubenswrapper[4492]: I0220 09:19:39.312528 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:19:39 crc kubenswrapper[4492]: I0220 09:19:39.742195 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtr54" event={"ID":"f4c8b933-df50-49a0-8c5f-f2d905a7be1c","Type":"ContainerStarted","Data":"72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1"} Feb 20 09:19:40 crc kubenswrapper[4492]: I0220 09:19:40.751696 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerID="72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1" exitCode=0 Feb 20 09:19:40 crc kubenswrapper[4492]: I0220 09:19:40.752001 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtr54" event={"ID":"f4c8b933-df50-49a0-8c5f-f2d905a7be1c","Type":"ContainerDied","Data":"72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1"} Feb 20 09:19:41 crc kubenswrapper[4492]: I0220 09:19:41.765756 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtr54" event={"ID":"f4c8b933-df50-49a0-8c5f-f2d905a7be1c","Type":"ContainerStarted","Data":"6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b"} Feb 20 09:19:41 crc kubenswrapper[4492]: I0220 09:19:41.799332 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jtr54" podStartSLOduration=4.151509818 podStartE2EDuration="6.797982547s" podCreationTimestamp="2026-02-20 09:19:35 +0000 UTC" firstStartedPulling="2026-02-20 09:19:38.731991899 +0000 UTC m=+9535.503280877" lastFinishedPulling="2026-02-20 09:19:41.378464637 +0000 UTC m=+9538.149753606" observedRunningTime="2026-02-20 09:19:41.793824878 +0000 UTC m=+9538.565113856" watchObservedRunningTime="2026-02-20 09:19:41.797982547 +0000 UTC m=+9538.569271526" Feb 20 09:19:46 crc kubenswrapper[4492]: I0220 09:19:46.915249 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:46 crc kubenswrapper[4492]: I0220 09:19:46.915912 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:47 crc kubenswrapper[4492]: I0220 09:19:47.341559 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:47 crc kubenswrapper[4492]: I0220 09:19:47.867360 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:47 crc kubenswrapper[4492]: I0220 09:19:47.919670 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtr54"] Feb 20 09:19:49 crc kubenswrapper[4492]: I0220 09:19:49.840170 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jtr54" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="registry-server" containerID="cri-o://6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b" gracePeriod=2 Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.351342 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.410655 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srpnm\" (UniqueName: \"kubernetes.io/projected/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-kube-api-access-srpnm\") pod \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.410915 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-catalog-content\") pod \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.411059 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-utilities\") pod \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\" (UID: \"f4c8b933-df50-49a0-8c5f-f2d905a7be1c\") " Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.412839 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-utilities" (OuterVolumeSpecName: "utilities") pod "f4c8b933-df50-49a0-8c5f-f2d905a7be1c" (UID: "f4c8b933-df50-49a0-8c5f-f2d905a7be1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.419716 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-kube-api-access-srpnm" (OuterVolumeSpecName: "kube-api-access-srpnm") pod "f4c8b933-df50-49a0-8c5f-f2d905a7be1c" (UID: "f4c8b933-df50-49a0-8c5f-f2d905a7be1c"). InnerVolumeSpecName "kube-api-access-srpnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.450636 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4c8b933-df50-49a0-8c5f-f2d905a7be1c" (UID: "f4c8b933-df50-49a0-8c5f-f2d905a7be1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.514090 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.514396 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.514409 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srpnm\" (UniqueName: \"kubernetes.io/projected/f4c8b933-df50-49a0-8c5f-f2d905a7be1c-kube-api-access-srpnm\") on node \"crc\" DevicePath \"\"" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.850502 4492 generic.go:334] "Generic (PLEG): container finished" podID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerID="6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b" exitCode=0 Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.850569 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtr54" event={"ID":"f4c8b933-df50-49a0-8c5f-f2d905a7be1c","Type":"ContainerDied","Data":"6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b"} Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.850616 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jtr54" event={"ID":"f4c8b933-df50-49a0-8c5f-f2d905a7be1c","Type":"ContainerDied","Data":"e67b7b278a96ee55d66871920be1a47602848574c5597686fc77f2632ec70241"} Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.850630 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jtr54" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.851363 4492 scope.go:117] "RemoveContainer" containerID="6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.875334 4492 scope.go:117] "RemoveContainer" containerID="72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.910851 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jtr54"] Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.913243 4492 scope.go:117] "RemoveContainer" containerID="014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.925096 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jtr54"] Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.952683 4492 scope.go:117] "RemoveContainer" containerID="6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b" Feb 20 09:19:50 crc kubenswrapper[4492]: E0220 09:19:50.954510 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b\": container with ID starting with 6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b not found: ID does not exist" containerID="6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.955145 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b"} err="failed to get container status \"6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b\": rpc error: code = NotFound desc = could not find container \"6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b\": container with ID starting with 6cd969c8372e1cbf38abb976f5b1dd542cae90cdfde4d2c1b44c55e2038a0c4b not found: ID does not exist" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.955180 4492 scope.go:117] "RemoveContainer" containerID="72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1" Feb 20 09:19:50 crc kubenswrapper[4492]: E0220 09:19:50.955745 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1\": container with ID starting with 72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1 not found: ID does not exist" containerID="72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.955785 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1"} err="failed to get container status \"72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1\": rpc error: code = NotFound desc = could not find container \"72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1\": container with ID starting with 72045e3c156f5897e94f956be5b86d14b92d5400017dabede1a88a9c26cbf9d1 not found: ID does not exist" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.955810 4492 scope.go:117] "RemoveContainer" containerID="014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666" Feb 20 09:19:50 crc kubenswrapper[4492]: E0220 09:19:50.956181 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666\": container with ID starting with 014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666 not found: ID does not exist" containerID="014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666" Feb 20 09:19:50 crc kubenswrapper[4492]: I0220 09:19:50.956241 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666"} err="failed to get container status \"014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666\": rpc error: code = NotFound desc = could not find container \"014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666\": container with ID starting with 014f5f41a85f7c44ce3a7e69c9e5512e348362271c1cbcceddc21ba863db6666 not found: ID does not exist" Feb 20 09:19:51 crc kubenswrapper[4492]: I0220 09:19:51.568127 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" path="/var/lib/kubelet/pods/f4c8b933-df50-49a0-8c5f-f2d905a7be1c/volumes" Feb 20 09:20:09 crc kubenswrapper[4492]: I0220 09:20:09.311298 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:20:09 crc kubenswrapper[4492]: I0220 09:20:09.312204 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:20:39 crc kubenswrapper[4492]: I0220 09:20:39.311349 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:20:39 crc kubenswrapper[4492]: I0220 09:20:39.312025 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:20:39 crc kubenswrapper[4492]: I0220 09:20:39.312082 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 09:20:39 crc kubenswrapper[4492]: I0220 09:20:39.312808 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ff6c52a7e322eb6e1d90878291c8341550e11afc0ae66488e8cec367610b76a"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:20:39 crc kubenswrapper[4492]: I0220 09:20:39.312869 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://6ff6c52a7e322eb6e1d90878291c8341550e11afc0ae66488e8cec367610b76a" gracePeriod=600 Feb 20 09:20:40 crc kubenswrapper[4492]: I0220 09:20:40.306821 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="6ff6c52a7e322eb6e1d90878291c8341550e11afc0ae66488e8cec367610b76a" exitCode=0 Feb 20 09:20:40 crc kubenswrapper[4492]: I0220 09:20:40.307196 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"6ff6c52a7e322eb6e1d90878291c8341550e11afc0ae66488e8cec367610b76a"} Feb 20 09:20:40 crc kubenswrapper[4492]: I0220 09:20:40.307247 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a"} Feb 20 09:20:40 crc kubenswrapper[4492]: I0220 09:20:40.307273 4492 scope.go:117] "RemoveContainer" containerID="bcf774aba082daf3e007d90b2743eee904592a58bfe4c7483e2bd67e05933d1b" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.687596 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4hqjm"] Feb 20 09:20:43 crc kubenswrapper[4492]: E0220 09:20:43.689182 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="extract-content" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.689201 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="extract-content" Feb 20 09:20:43 crc kubenswrapper[4492]: E0220 09:20:43.689226 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="registry-server" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.689232 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="registry-server" Feb 20 09:20:43 crc kubenswrapper[4492]: E0220 09:20:43.689248 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="extract-utilities" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.689255 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="extract-utilities" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.689522 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4c8b933-df50-49a0-8c5f-f2d905a7be1c" containerName="registry-server" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.692678 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.722028 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-utilities\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.722177 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s7b7\" (UniqueName: \"kubernetes.io/projected/c32456b0-26b8-4835-91c4-7c26f646d3e9-kube-api-access-9s7b7\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.722689 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-catalog-content\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.724957 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hqjm"] Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.835378 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-catalog-content\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.835491 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-utilities\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.835538 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s7b7\" (UniqueName: \"kubernetes.io/projected/c32456b0-26b8-4835-91c4-7c26f646d3e9-kube-api-access-9s7b7\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.835830 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-catalog-content\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.836082 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-utilities\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:43 crc kubenswrapper[4492]: I0220 09:20:43.852888 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s7b7\" (UniqueName: \"kubernetes.io/projected/c32456b0-26b8-4835-91c4-7c26f646d3e9-kube-api-access-9s7b7\") pod \"community-operators-4hqjm\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:44 crc kubenswrapper[4492]: I0220 09:20:44.028294 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:44 crc kubenswrapper[4492]: I0220 09:20:44.527315 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hqjm"] Feb 20 09:20:44 crc kubenswrapper[4492]: W0220 09:20:44.542307 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc32456b0_26b8_4835_91c4_7c26f646d3e9.slice/crio-162232c1b6f68fe245777e5a8e5f593336db5bdf87d3afceb122e571d64fbac4 WatchSource:0}: Error finding container 162232c1b6f68fe245777e5a8e5f593336db5bdf87d3afceb122e571d64fbac4: Status 404 returned error can't find the container with id 162232c1b6f68fe245777e5a8e5f593336db5bdf87d3afceb122e571d64fbac4 Feb 20 09:20:45 crc kubenswrapper[4492]: I0220 09:20:45.361576 4492 generic.go:334] "Generic (PLEG): container finished" podID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerID="8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474" exitCode=0 Feb 20 09:20:45 crc kubenswrapper[4492]: I0220 09:20:45.361975 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hqjm" event={"ID":"c32456b0-26b8-4835-91c4-7c26f646d3e9","Type":"ContainerDied","Data":"8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474"} Feb 20 09:20:45 crc kubenswrapper[4492]: I0220 09:20:45.362022 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hqjm" event={"ID":"c32456b0-26b8-4835-91c4-7c26f646d3e9","Type":"ContainerStarted","Data":"162232c1b6f68fe245777e5a8e5f593336db5bdf87d3afceb122e571d64fbac4"} Feb 20 09:20:46 crc kubenswrapper[4492]: I0220 09:20:46.373751 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hqjm" event={"ID":"c32456b0-26b8-4835-91c4-7c26f646d3e9","Type":"ContainerStarted","Data":"dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260"} Feb 20 09:20:46 crc kubenswrapper[4492]: I0220 09:20:46.831174 4492 oom_watcher_linux.go:83] "Got sys oom event" event={"Pid":283835,"ProcessName":"python3","TimeOfDeath":"2026-02-20T09:20:46.644499866Z","ContainerName":"/","VictimContainerName":"/","Constraint":""} Feb 20 09:20:47 crc kubenswrapper[4492]: I0220 09:20:47.387746 4492 generic.go:334] "Generic (PLEG): container finished" podID="0d1fd559-3c53-44c1-aa9b-556b306a479b" containerID="33b4912328c192d6e2377bb42694cffe83780ea6af9e48234e0fc59923cd57db" exitCode=137 Feb 20 09:20:47 crc kubenswrapper[4492]: I0220 09:20:47.387808 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"0d1fd559-3c53-44c1-aa9b-556b306a479b","Type":"ContainerDied","Data":"33b4912328c192d6e2377bb42694cffe83780ea6af9e48234e0fc59923cd57db"} Feb 20 09:20:47 crc kubenswrapper[4492]: I0220 09:20:47.390154 4492 generic.go:334] "Generic (PLEG): container finished" podID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerID="dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260" exitCode=0 Feb 20 09:20:47 crc kubenswrapper[4492]: I0220 09:20:47.390193 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hqjm" event={"ID":"c32456b0-26b8-4835-91c4-7c26f646d3e9","Type":"ContainerDied","Data":"dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260"} Feb 20 09:20:48 crc kubenswrapper[4492]: I0220 09:20:48.403161 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hqjm" event={"ID":"c32456b0-26b8-4835-91c4-7c26f646d3e9","Type":"ContainerStarted","Data":"559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845"} Feb 20 09:20:48 crc kubenswrapper[4492]: I0220 09:20:48.426033 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4hqjm" podStartSLOduration=2.946435223 podStartE2EDuration="5.426007299s" podCreationTimestamp="2026-02-20 09:20:43 +0000 UTC" firstStartedPulling="2026-02-20 09:20:45.366297863 +0000 UTC m=+9602.137586842" lastFinishedPulling="2026-02-20 09:20:47.845869939 +0000 UTC m=+9604.617158918" observedRunningTime="2026-02-20 09:20:48.424908988 +0000 UTC m=+9605.196197966" watchObservedRunningTime="2026-02-20 09:20:48.426007299 +0000 UTC m=+9605.197296277" Feb 20 09:20:48 crc kubenswrapper[4492]: I0220 09:20:48.983661 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068227 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ca-certs\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068345 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config-secret\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068427 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-config-data\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068494 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ssh-key\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068602 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-workdir\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068694 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068735 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-temporary\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068796 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.068921 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dpsg\" (UniqueName: \"kubernetes.io/projected/0d1fd559-3c53-44c1-aa9b-556b306a479b-kube-api-access-6dpsg\") pod \"0d1fd559-3c53-44c1-aa9b-556b306a479b\" (UID: \"0d1fd559-3c53-44c1-aa9b-556b306a479b\") " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.069840 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.073411 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.074643 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-config-data" (OuterVolumeSpecName: "config-data") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.094940 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d1fd559-3c53-44c1-aa9b-556b306a479b-kube-api-access-6dpsg" (OuterVolumeSpecName: "kube-api-access-6dpsg") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "kube-api-access-6dpsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.094924 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.100954 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.101342 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.107065 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.119050 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "0d1fd559-3c53-44c1-aa9b-556b306a479b" (UID: "0d1fd559-3c53-44c1-aa9b-556b306a479b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172482 4492 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172525 4492 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172545 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172559 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dpsg\" (UniqueName: \"kubernetes.io/projected/0d1fd559-3c53-44c1-aa9b-556b306a479b-kube-api-access-6dpsg\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172570 4492 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172581 4492 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172591 4492 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0d1fd559-3c53-44c1-aa9b-556b306a479b-config-data\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172605 4492 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0d1fd559-3c53-44c1-aa9b-556b306a479b-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.172615 4492 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/0d1fd559-3c53-44c1-aa9b-556b306a479b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.188588 4492 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.274435 4492 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.413615 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"0d1fd559-3c53-44c1-aa9b-556b306a479b","Type":"ContainerDied","Data":"1afa41b1fe42bc6ab6c4d34bdd4030962a3db49ebd170cf869e7575aa8fca2a7"} Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.413692 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1afa41b1fe42bc6ab6c4d34bdd4030962a3db49ebd170cf869e7575aa8fca2a7" Feb 20 09:20:49 crc kubenswrapper[4492]: I0220 09:20:49.413649 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Feb 20 09:20:54 crc kubenswrapper[4492]: I0220 09:20:54.029935 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:54 crc kubenswrapper[4492]: I0220 09:20:54.030672 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:54 crc kubenswrapper[4492]: I0220 09:20:54.067189 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:54 crc kubenswrapper[4492]: I0220 09:20:54.490893 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:54 crc kubenswrapper[4492]: I0220 09:20:54.532432 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hqjm"] Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.478266 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4hqjm" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="registry-server" containerID="cri-o://559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845" gracePeriod=2 Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.872605 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.916858 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.917017 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s7b7\" (UniqueName: \"kubernetes.io/projected/c32456b0-26b8-4835-91c4-7c26f646d3e9-kube-api-access-9s7b7\") pod \"c32456b0-26b8-4835-91c4-7c26f646d3e9\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.917164 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-catalog-content\") pod \"c32456b0-26b8-4835-91c4-7c26f646d3e9\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.917569 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-utilities\") pod \"c32456b0-26b8-4835-91c4-7c26f646d3e9\" (UID: \"c32456b0-26b8-4835-91c4-7c26f646d3e9\") " Feb 20 09:20:56 crc kubenswrapper[4492]: E0220 09:20:56.917859 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="extract-content" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.917885 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="extract-content" Feb 20 09:20:56 crc kubenswrapper[4492]: E0220 09:20:56.917905 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="extract-utilities" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.917914 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="extract-utilities" Feb 20 09:20:56 crc kubenswrapper[4492]: E0220 09:20:56.917931 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d1fd559-3c53-44c1-aa9b-556b306a479b" containerName="tempest-tests-tempest-tests-runner" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.917938 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d1fd559-3c53-44c1-aa9b-556b306a479b" containerName="tempest-tests-tempest-tests-runner" Feb 20 09:20:56 crc kubenswrapper[4492]: E0220 09:20:56.917959 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="registry-server" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.917964 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="registry-server" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.918156 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerName="registry-server" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.918180 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d1fd559-3c53-44c1-aa9b-556b306a479b" containerName="tempest-tests-tempest-tests-runner" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.919009 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-utilities" (OuterVolumeSpecName: "utilities") pod "c32456b0-26b8-4835-91c4-7c26f646d3e9" (UID: "c32456b0-26b8-4835-91c4-7c26f646d3e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.919206 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.925720 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c32456b0-26b8-4835-91c4-7c26f646d3e9-kube-api-access-9s7b7" (OuterVolumeSpecName: "kube-api-access-9s7b7") pod "c32456b0-26b8-4835-91c4-7c26f646d3e9" (UID: "c32456b0-26b8-4835-91c4-7c26f646d3e9"). InnerVolumeSpecName "kube-api-access-9s7b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.930689 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-pdp6n" Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.949535 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 09:20:56 crc kubenswrapper[4492]: I0220 09:20:56.971623 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c32456b0-26b8-4835-91c4-7c26f646d3e9" (UID: "c32456b0-26b8-4835-91c4-7c26f646d3e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.020542 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2ae49fdb-33f1-44a8-a26f-222c204cdfac\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.020670 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhsgs\" (UniqueName: \"kubernetes.io/projected/2ae49fdb-33f1-44a8-a26f-222c204cdfac-kube-api-access-lhsgs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2ae49fdb-33f1-44a8-a26f-222c204cdfac\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.020770 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.020793 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s7b7\" (UniqueName: \"kubernetes.io/projected/c32456b0-26b8-4835-91c4-7c26f646d3e9-kube-api-access-9s7b7\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.020802 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c32456b0-26b8-4835-91c4-7c26f646d3e9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.122411 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2ae49fdb-33f1-44a8-a26f-222c204cdfac\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.122560 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhsgs\" (UniqueName: \"kubernetes.io/projected/2ae49fdb-33f1-44a8-a26f-222c204cdfac-kube-api-access-lhsgs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2ae49fdb-33f1-44a8-a26f-222c204cdfac\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.123886 4492 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2ae49fdb-33f1-44a8-a26f-222c204cdfac\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.138238 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhsgs\" (UniqueName: \"kubernetes.io/projected/2ae49fdb-33f1-44a8-a26f-222c204cdfac-kube-api-access-lhsgs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2ae49fdb-33f1-44a8-a26f-222c204cdfac\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.146651 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"2ae49fdb-33f1-44a8-a26f-222c204cdfac\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.277708 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.490401 4492 generic.go:334] "Generic (PLEG): container finished" podID="c32456b0-26b8-4835-91c4-7c26f646d3e9" containerID="559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845" exitCode=0 Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.490489 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hqjm" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.490517 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hqjm" event={"ID":"c32456b0-26b8-4835-91c4-7c26f646d3e9","Type":"ContainerDied","Data":"559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845"} Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.490832 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hqjm" event={"ID":"c32456b0-26b8-4835-91c4-7c26f646d3e9","Type":"ContainerDied","Data":"162232c1b6f68fe245777e5a8e5f593336db5bdf87d3afceb122e571d64fbac4"} Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.490852 4492 scope.go:117] "RemoveContainer" containerID="559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.512198 4492 scope.go:117] "RemoveContainer" containerID="dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.531593 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hqjm"] Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.542467 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4hqjm"] Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.548772 4492 scope.go:117] "RemoveContainer" containerID="8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.568568 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c32456b0-26b8-4835-91c4-7c26f646d3e9" path="/var/lib/kubelet/pods/c32456b0-26b8-4835-91c4-7c26f646d3e9/volumes" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.571580 4492 scope.go:117] "RemoveContainer" containerID="559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845" Feb 20 09:20:57 crc kubenswrapper[4492]: E0220 09:20:57.572155 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845\": container with ID starting with 559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845 not found: ID does not exist" containerID="559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.572193 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845"} err="failed to get container status \"559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845\": rpc error: code = NotFound desc = could not find container \"559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845\": container with ID starting with 559ba830ee71ecbe2038bfcc14032b1857a37b2efdbea6cd0f2a0ee26f6be845 not found: ID does not exist" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.572213 4492 scope.go:117] "RemoveContainer" containerID="dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260" Feb 20 09:20:57 crc kubenswrapper[4492]: E0220 09:20:57.572533 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260\": container with ID starting with dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260 not found: ID does not exist" containerID="dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.572554 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260"} err="failed to get container status \"dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260\": rpc error: code = NotFound desc = could not find container \"dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260\": container with ID starting with dd4fbe4b3e879499cf30b6b27439bbf5c1e817d898ce50cc829461b82c9f8260 not found: ID does not exist" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.572567 4492 scope.go:117] "RemoveContainer" containerID="8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474" Feb 20 09:20:57 crc kubenswrapper[4492]: E0220 09:20:57.572862 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474\": container with ID starting with 8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474 not found: ID does not exist" containerID="8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.572882 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474"} err="failed to get container status \"8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474\": rpc error: code = NotFound desc = could not find container \"8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474\": container with ID starting with 8a0eeadc2084f8bcdb0f695921875d91536a9dbb519949e9c52bceb9a70a5474 not found: ID does not exist" Feb 20 09:20:57 crc kubenswrapper[4492]: I0220 09:20:57.682402 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 20 09:20:58 crc kubenswrapper[4492]: I0220 09:20:58.500714 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2ae49fdb-33f1-44a8-a26f-222c204cdfac","Type":"ContainerStarted","Data":"f6ee5ea1c6bc2d36b81fbd726c56404412408e820e9f13838ed4c0f168a4ba7d"} Feb 20 09:20:59 crc kubenswrapper[4492]: I0220 09:20:59.512437 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"2ae49fdb-33f1-44a8-a26f-222c204cdfac","Type":"ContainerStarted","Data":"3c41f9bfed7509df271a0d77f779388a6eec8d1ed71fe041be9dffb45a8e0f07"} Feb 20 09:20:59 crc kubenswrapper[4492]: I0220 09:20:59.535120 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.622957477 podStartE2EDuration="3.535100396s" podCreationTimestamp="2026-02-20 09:20:56 +0000 UTC" firstStartedPulling="2026-02-20 09:20:57.691013521 +0000 UTC m=+9614.462302499" lastFinishedPulling="2026-02-20 09:20:58.60315644 +0000 UTC m=+9615.374445418" observedRunningTime="2026-02-20 09:20:59.524826724 +0000 UTC m=+9616.296115701" watchObservedRunningTime="2026-02-20 09:20:59.535100396 +0000 UTC m=+9616.306389374" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.598484 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7grgm/must-gather-rdc6g"] Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.606142 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.613391 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7grgm"/"openshift-service-ca.crt" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.614648 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7grgm"/"kube-root-ca.crt" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.634097 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7grgm/must-gather-rdc6g"] Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.748555 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sqt7\" (UniqueName: \"kubernetes.io/projected/37384967-b137-4cdf-868d-74cb7dfcc85c-kube-api-access-8sqt7\") pod \"must-gather-rdc6g\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.748619 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/37384967-b137-4cdf-868d-74cb7dfcc85c-must-gather-output\") pod \"must-gather-rdc6g\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.851416 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sqt7\" (UniqueName: \"kubernetes.io/projected/37384967-b137-4cdf-868d-74cb7dfcc85c-kube-api-access-8sqt7\") pod \"must-gather-rdc6g\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.851525 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/37384967-b137-4cdf-868d-74cb7dfcc85c-must-gather-output\") pod \"must-gather-rdc6g\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.852061 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/37384967-b137-4cdf-868d-74cb7dfcc85c-must-gather-output\") pod \"must-gather-rdc6g\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.869907 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sqt7\" (UniqueName: \"kubernetes.io/projected/37384967-b137-4cdf-868d-74cb7dfcc85c-kube-api-access-8sqt7\") pod \"must-gather-rdc6g\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:30 crc kubenswrapper[4492]: I0220 09:21:30.924390 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:21:31 crc kubenswrapper[4492]: I0220 09:21:31.457457 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7grgm/must-gather-rdc6g"] Feb 20 09:21:31 crc kubenswrapper[4492]: I0220 09:21:31.823688 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/must-gather-rdc6g" event={"ID":"37384967-b137-4cdf-868d-74cb7dfcc85c","Type":"ContainerStarted","Data":"e879f94f1c8e7406d9e0fc6e927f9f15678a997a5a7bf156af9bedd6474a792c"} Feb 20 09:21:39 crc kubenswrapper[4492]: I0220 09:21:39.937249 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/must-gather-rdc6g" event={"ID":"37384967-b137-4cdf-868d-74cb7dfcc85c","Type":"ContainerStarted","Data":"358fec6ee7a399344302d202cfd505b965459adc45196c72c3ba135472c18b65"} Feb 20 09:21:39 crc kubenswrapper[4492]: I0220 09:21:39.937891 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/must-gather-rdc6g" event={"ID":"37384967-b137-4cdf-868d-74cb7dfcc85c","Type":"ContainerStarted","Data":"7029bca25b42ac67a0250597d1aaab172eee7f40f68b5548fdc3ad9f759c23cf"} Feb 20 09:21:39 crc kubenswrapper[4492]: I0220 09:21:39.966310 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7grgm/must-gather-rdc6g" podStartSLOduration=2.224354232 podStartE2EDuration="9.966282556s" podCreationTimestamp="2026-02-20 09:21:30 +0000 UTC" firstStartedPulling="2026-02-20 09:21:31.496142608 +0000 UTC m=+9648.267431587" lastFinishedPulling="2026-02-20 09:21:39.238070933 +0000 UTC m=+9656.009359911" observedRunningTime="2026-02-20 09:21:39.959446387 +0000 UTC m=+9656.730735365" watchObservedRunningTime="2026-02-20 09:21:39.966282556 +0000 UTC m=+9656.737571534" Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.590337 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7grgm/crc-debug-w2bw9"] Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.593065 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.595653 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7grgm"/"default-dockercfg-zx4jv" Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.659872 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spmtt\" (UniqueName: \"kubernetes.io/projected/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-kube-api-access-spmtt\") pod \"crc-debug-w2bw9\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.659982 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-host\") pod \"crc-debug-w2bw9\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.762435 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spmtt\" (UniqueName: \"kubernetes.io/projected/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-kube-api-access-spmtt\") pod \"crc-debug-w2bw9\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.762569 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-host\") pod \"crc-debug-w2bw9\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:43 crc kubenswrapper[4492]: I0220 09:21:43.763408 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-host\") pod \"crc-debug-w2bw9\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:44 crc kubenswrapper[4492]: I0220 09:21:44.120251 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spmtt\" (UniqueName: \"kubernetes.io/projected/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-kube-api-access-spmtt\") pod \"crc-debug-w2bw9\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:44 crc kubenswrapper[4492]: I0220 09:21:44.220810 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7grgm"/"default-dockercfg-zx4jv" Feb 20 09:21:44 crc kubenswrapper[4492]: I0220 09:21:44.230011 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:21:44 crc kubenswrapper[4492]: I0220 09:21:44.979739 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" event={"ID":"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed","Type":"ContainerStarted","Data":"ba713036eac3f594517bc632e69605fac85f3dab4a6ee244167c8ff525e17c8b"} Feb 20 09:21:56 crc kubenswrapper[4492]: I0220 09:21:56.120515 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" event={"ID":"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed","Type":"ContainerStarted","Data":"f78c0eb0a607de08007736cd2f9ad301dcd228413bf9a1a39588a24f365eb08f"} Feb 20 09:21:56 crc kubenswrapper[4492]: I0220 09:21:56.138368 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" podStartSLOduration=2.213069699 podStartE2EDuration="13.138346448s" podCreationTimestamp="2026-02-20 09:21:43 +0000 UTC" firstStartedPulling="2026-02-20 09:21:44.301596793 +0000 UTC m=+9661.072885771" lastFinishedPulling="2026-02-20 09:21:55.226873542 +0000 UTC m=+9671.998162520" observedRunningTime="2026-02-20 09:21:56.136042433 +0000 UTC m=+9672.907331412" watchObservedRunningTime="2026-02-20 09:21:56.138346448 +0000 UTC m=+9672.909635425" Feb 20 09:22:39 crc kubenswrapper[4492]: I0220 09:22:39.312260 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:22:39 crc kubenswrapper[4492]: I0220 09:22:39.313048 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:22:42 crc kubenswrapper[4492]: I0220 09:22:42.557119 4492 generic.go:334] "Generic (PLEG): container finished" podID="84643317-76f3-4cc5-ab98-bf8bf9cbf3ed" containerID="f78c0eb0a607de08007736cd2f9ad301dcd228413bf9a1a39588a24f365eb08f" exitCode=0 Feb 20 09:22:42 crc kubenswrapper[4492]: I0220 09:22:42.557338 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" event={"ID":"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed","Type":"ContainerDied","Data":"f78c0eb0a607de08007736cd2f9ad301dcd228413bf9a1a39588a24f365eb08f"} Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.697729 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.732716 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7grgm/crc-debug-w2bw9"] Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.743214 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7grgm/crc-debug-w2bw9"] Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.858834 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-host\") pod \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.859067 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spmtt\" (UniqueName: \"kubernetes.io/projected/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-kube-api-access-spmtt\") pod \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\" (UID: \"84643317-76f3-4cc5-ab98-bf8bf9cbf3ed\") " Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.860607 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-host" (OuterVolumeSpecName: "host") pod "84643317-76f3-4cc5-ab98-bf8bf9cbf3ed" (UID: "84643317-76f3-4cc5-ab98-bf8bf9cbf3ed"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.866891 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-kube-api-access-spmtt" (OuterVolumeSpecName: "kube-api-access-spmtt") pod "84643317-76f3-4cc5-ab98-bf8bf9cbf3ed" (UID: "84643317-76f3-4cc5-ab98-bf8bf9cbf3ed"). InnerVolumeSpecName "kube-api-access-spmtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.962586 4492 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:22:43 crc kubenswrapper[4492]: I0220 09:22:43.962616 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spmtt\" (UniqueName: \"kubernetes.io/projected/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed-kube-api-access-spmtt\") on node \"crc\" DevicePath \"\"" Feb 20 09:22:44 crc kubenswrapper[4492]: I0220 09:22:44.583451 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba713036eac3f594517bc632e69605fac85f3dab4a6ee244167c8ff525e17c8b" Feb 20 09:22:44 crc kubenswrapper[4492]: I0220 09:22:44.583522 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-w2bw9" Feb 20 09:22:44 crc kubenswrapper[4492]: I0220 09:22:44.925436 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7grgm/crc-debug-9m58j"] Feb 20 09:22:44 crc kubenswrapper[4492]: E0220 09:22:44.926908 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84643317-76f3-4cc5-ab98-bf8bf9cbf3ed" containerName="container-00" Feb 20 09:22:44 crc kubenswrapper[4492]: I0220 09:22:44.926970 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="84643317-76f3-4cc5-ab98-bf8bf9cbf3ed" containerName="container-00" Feb 20 09:22:44 crc kubenswrapper[4492]: I0220 09:22:44.927770 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="84643317-76f3-4cc5-ab98-bf8bf9cbf3ed" containerName="container-00" Feb 20 09:22:44 crc kubenswrapper[4492]: I0220 09:22:44.929309 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:44 crc kubenswrapper[4492]: I0220 09:22:44.932308 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7grgm"/"default-dockercfg-zx4jv" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.092150 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l64tk\" (UniqueName: \"kubernetes.io/projected/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-kube-api-access-l64tk\") pod \"crc-debug-9m58j\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.092541 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-host\") pod \"crc-debug-9m58j\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.193687 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l64tk\" (UniqueName: \"kubernetes.io/projected/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-kube-api-access-l64tk\") pod \"crc-debug-9m58j\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.193726 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-host\") pod \"crc-debug-9m58j\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.193957 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-host\") pod \"crc-debug-9m58j\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.212002 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l64tk\" (UniqueName: \"kubernetes.io/projected/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-kube-api-access-l64tk\") pod \"crc-debug-9m58j\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.251991 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:45 crc kubenswrapper[4492]: W0220 09:22:45.302985 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e7e5e5c_d4ff_4792_bb6f_50541fbb5f4e.slice/crio-e3b8f6242992262a959d360f29072edac4b16bd6589261ff90516f36fbcb7354 WatchSource:0}: Error finding container e3b8f6242992262a959d360f29072edac4b16bd6589261ff90516f36fbcb7354: Status 404 returned error can't find the container with id e3b8f6242992262a959d360f29072edac4b16bd6589261ff90516f36fbcb7354 Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.571242 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84643317-76f3-4cc5-ab98-bf8bf9cbf3ed" path="/var/lib/kubelet/pods/84643317-76f3-4cc5-ab98-bf8bf9cbf3ed/volumes" Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.602525 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-9m58j" event={"ID":"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e","Type":"ContainerStarted","Data":"2d7dd333ce0014f99fb6ef295945aacd68033f536f4f8ae56477187165135221"} Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.602604 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-9m58j" event={"ID":"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e","Type":"ContainerStarted","Data":"e3b8f6242992262a959d360f29072edac4b16bd6589261ff90516f36fbcb7354"} Feb 20 09:22:45 crc kubenswrapper[4492]: I0220 09:22:45.625283 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7grgm/crc-debug-9m58j" podStartSLOduration=1.625266725 podStartE2EDuration="1.625266725s" podCreationTimestamp="2026-02-20 09:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-20 09:22:45.618546764 +0000 UTC m=+9722.389835742" watchObservedRunningTime="2026-02-20 09:22:45.625266725 +0000 UTC m=+9722.396555703" Feb 20 09:22:46 crc kubenswrapper[4492]: I0220 09:22:46.615888 4492 generic.go:334] "Generic (PLEG): container finished" podID="7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e" containerID="2d7dd333ce0014f99fb6ef295945aacd68033f536f4f8ae56477187165135221" exitCode=0 Feb 20 09:22:46 crc kubenswrapper[4492]: I0220 09:22:46.615936 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-9m58j" event={"ID":"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e","Type":"ContainerDied","Data":"2d7dd333ce0014f99fb6ef295945aacd68033f536f4f8ae56477187165135221"} Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.711342 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.751261 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7grgm/crc-debug-9m58j"] Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.763635 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7grgm/crc-debug-9m58j"] Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.875966 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l64tk\" (UniqueName: \"kubernetes.io/projected/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-kube-api-access-l64tk\") pod \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.876289 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-host\") pod \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\" (UID: \"7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e\") " Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.876420 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-host" (OuterVolumeSpecName: "host") pod "7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e" (UID: "7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.877209 4492 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.881908 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-kube-api-access-l64tk" (OuterVolumeSpecName: "kube-api-access-l64tk") pod "7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e" (UID: "7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e"). InnerVolumeSpecName "kube-api-access-l64tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:22:47 crc kubenswrapper[4492]: I0220 09:22:47.980568 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l64tk\" (UniqueName: \"kubernetes.io/projected/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e-kube-api-access-l64tk\") on node \"crc\" DevicePath \"\"" Feb 20 09:22:48 crc kubenswrapper[4492]: I0220 09:22:48.636369 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3b8f6242992262a959d360f29072edac4b16bd6589261ff90516f36fbcb7354" Feb 20 09:22:48 crc kubenswrapper[4492]: I0220 09:22:48.636454 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-9m58j" Feb 20 09:22:48 crc kubenswrapper[4492]: I0220 09:22:48.954978 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7grgm/crc-debug-jtvkd"] Feb 20 09:22:48 crc kubenswrapper[4492]: E0220 09:22:48.957038 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e" containerName="container-00" Feb 20 09:22:48 crc kubenswrapper[4492]: I0220 09:22:48.957187 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e" containerName="container-00" Feb 20 09:22:48 crc kubenswrapper[4492]: I0220 09:22:48.957562 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e" containerName="container-00" Feb 20 09:22:48 crc kubenswrapper[4492]: I0220 09:22:48.958685 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:48 crc kubenswrapper[4492]: I0220 09:22:48.962009 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7grgm"/"default-dockercfg-zx4jv" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.109613 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-host\") pod \"crc-debug-jtvkd\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.109727 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgqzf\" (UniqueName: \"kubernetes.io/projected/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-kube-api-access-jgqzf\") pod \"crc-debug-jtvkd\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.211599 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgqzf\" (UniqueName: \"kubernetes.io/projected/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-kube-api-access-jgqzf\") pod \"crc-debug-jtvkd\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.211804 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-host\") pod \"crc-debug-jtvkd\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.211910 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-host\") pod \"crc-debug-jtvkd\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.230622 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgqzf\" (UniqueName: \"kubernetes.io/projected/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-kube-api-access-jgqzf\") pod \"crc-debug-jtvkd\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.281144 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.573501 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e" path="/var/lib/kubelet/pods/7e7e5e5c-d4ff-4792-bb6f-50541fbb5f4e/volumes" Feb 20 09:22:49 crc kubenswrapper[4492]: I0220 09:22:49.650416 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-jtvkd" event={"ID":"40b6c08e-ba49-4b3d-a24e-aadf94ba8244","Type":"ContainerStarted","Data":"67b51adc601be2b299dec83861e19fc696a688486e79d7aeead996c84078cf04"} Feb 20 09:22:50 crc kubenswrapper[4492]: I0220 09:22:50.661546 4492 generic.go:334] "Generic (PLEG): container finished" podID="40b6c08e-ba49-4b3d-a24e-aadf94ba8244" containerID="0d6fce88426c4c0bc2d1af55846da8a488a16e8560ea99fa7fdf2d0b352fd968" exitCode=0 Feb 20 09:22:50 crc kubenswrapper[4492]: I0220 09:22:50.661758 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/crc-debug-jtvkd" event={"ID":"40b6c08e-ba49-4b3d-a24e-aadf94ba8244","Type":"ContainerDied","Data":"0d6fce88426c4c0bc2d1af55846da8a488a16e8560ea99fa7fdf2d0b352fd968"} Feb 20 09:22:50 crc kubenswrapper[4492]: I0220 09:22:50.718607 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7grgm/crc-debug-jtvkd"] Feb 20 09:22:50 crc kubenswrapper[4492]: I0220 09:22:50.726824 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7grgm/crc-debug-jtvkd"] Feb 20 09:22:51 crc kubenswrapper[4492]: I0220 09:22:51.752096 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:51 crc kubenswrapper[4492]: I0220 09:22:51.873303 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgqzf\" (UniqueName: \"kubernetes.io/projected/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-kube-api-access-jgqzf\") pod \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " Feb 20 09:22:51 crc kubenswrapper[4492]: I0220 09:22:51.873639 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-host\") pod \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\" (UID: \"40b6c08e-ba49-4b3d-a24e-aadf94ba8244\") " Feb 20 09:22:51 crc kubenswrapper[4492]: I0220 09:22:51.873780 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-host" (OuterVolumeSpecName: "host") pod "40b6c08e-ba49-4b3d-a24e-aadf94ba8244" (UID: "40b6c08e-ba49-4b3d-a24e-aadf94ba8244"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 20 09:22:51 crc kubenswrapper[4492]: I0220 09:22:51.874254 4492 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-host\") on node \"crc\" DevicePath \"\"" Feb 20 09:22:51 crc kubenswrapper[4492]: I0220 09:22:51.883342 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-kube-api-access-jgqzf" (OuterVolumeSpecName: "kube-api-access-jgqzf") pod "40b6c08e-ba49-4b3d-a24e-aadf94ba8244" (UID: "40b6c08e-ba49-4b3d-a24e-aadf94ba8244"). InnerVolumeSpecName "kube-api-access-jgqzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:22:51 crc kubenswrapper[4492]: I0220 09:22:51.976948 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgqzf\" (UniqueName: \"kubernetes.io/projected/40b6c08e-ba49-4b3d-a24e-aadf94ba8244-kube-api-access-jgqzf\") on node \"crc\" DevicePath \"\"" Feb 20 09:22:52 crc kubenswrapper[4492]: I0220 09:22:52.685448 4492 scope.go:117] "RemoveContainer" containerID="0d6fce88426c4c0bc2d1af55846da8a488a16e8560ea99fa7fdf2d0b352fd968" Feb 20 09:22:52 crc kubenswrapper[4492]: I0220 09:22:52.685546 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/crc-debug-jtvkd" Feb 20 09:22:53 crc kubenswrapper[4492]: I0220 09:22:53.572677 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40b6c08e-ba49-4b3d-a24e-aadf94ba8244" path="/var/lib/kubelet/pods/40b6c08e-ba49-4b3d-a24e-aadf94ba8244/volumes" Feb 20 09:23:09 crc kubenswrapper[4492]: I0220 09:23:09.311555 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:23:09 crc kubenswrapper[4492]: I0220 09:23:09.312417 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:23:15 crc kubenswrapper[4492]: I0220 09:23:15.028535 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9db58c4f4-5prd7_4e3149f8-a4fe-484c-a0eb-9799994a144c/barbican-api/0.log" Feb 20 09:23:15 crc kubenswrapper[4492]: I0220 09:23:15.152445 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9db58c4f4-5prd7_4e3149f8-a4fe-484c-a0eb-9799994a144c/barbican-api-log/0.log" Feb 20 09:23:15 crc kubenswrapper[4492]: I0220 09:23:15.309594 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b75bfd866-zx65r_fa37a18e-7d2d-45db-8988-0ec443201155/barbican-keystone-listener/0.log" Feb 20 09:23:15 crc kubenswrapper[4492]: I0220 09:23:15.440580 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7b75bfd866-zx65r_fa37a18e-7d2d-45db-8988-0ec443201155/barbican-keystone-listener-log/0.log" Feb 20 09:23:15 crc kubenswrapper[4492]: I0220 09:23:15.788630 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f5cc678f7-42jkm_2d3f17e6-f16a-46db-a73c-331307b540cd/barbican-worker/0.log" Feb 20 09:23:15 crc kubenswrapper[4492]: I0220 09:23:15.826615 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-f5cc678f7-42jkm_2d3f17e6-f16a-46db-a73c-331307b540cd/barbican-worker-log/0.log" Feb 20 09:23:15 crc kubenswrapper[4492]: I0220 09:23:15.934627 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-5vql4_52acdbf4-c839-4ed1-bbd7-a7674e2bab77/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:16 crc kubenswrapper[4492]: I0220 09:23:16.318154 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0e1abd4-7d2e-493b-b545-180dce022a33/ceilometer-central-agent/0.log" Feb 20 09:23:16 crc kubenswrapper[4492]: I0220 09:23:16.399790 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0e1abd4-7d2e-493b-b545-180dce022a33/proxy-httpd/0.log" Feb 20 09:23:16 crc kubenswrapper[4492]: I0220 09:23:16.447809 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0e1abd4-7d2e-493b-b545-180dce022a33/ceilometer-notification-agent/0.log" Feb 20 09:23:16 crc kubenswrapper[4492]: I0220 09:23:16.640925 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0e1abd4-7d2e-493b-b545-180dce022a33/sg-core/0.log" Feb 20 09:23:16 crc kubenswrapper[4492]: I0220 09:23:16.671573 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_86f25515-0393-402b-a3f9-1c787e7e911e/cinder-api-log/0.log" Feb 20 09:23:16 crc kubenswrapper[4492]: I0220 09:23:16.785923 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_86f25515-0393-402b-a3f9-1c787e7e911e/cinder-api/0.log" Feb 20 09:23:16 crc kubenswrapper[4492]: I0220 09:23:16.916959 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_343b4ca3-ca66-42d1-9b10-8c5d5c9863e5/cinder-scheduler/0.log" Feb 20 09:23:17 crc kubenswrapper[4492]: I0220 09:23:17.035862 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_343b4ca3-ca66-42d1-9b10-8c5d5c9863e5/probe/0.log" Feb 20 09:23:17 crc kubenswrapper[4492]: I0220 09:23:17.204580 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-2m7tj_78fa6f40-87e3-4c44-b233-0a83c8e8d429/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:17 crc kubenswrapper[4492]: I0220 09:23:17.809724 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7c759b9c5-fr6n5_b70a74af-cdf1-4c6a-92ce-1148516ae7ee/init/0.log" Feb 20 09:23:17 crc kubenswrapper[4492]: I0220 09:23:17.854492 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-dcvxd_719a407b-3228-4bbc-a600-c08f555ba556/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:17 crc kubenswrapper[4492]: I0220 09:23:17.986136 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7c759b9c5-fr6n5_b70a74af-cdf1-4c6a-92ce-1148516ae7ee/init/0.log" Feb 20 09:23:18 crc kubenswrapper[4492]: I0220 09:23:18.118411 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-gr2zr_d0c74792-a006-4752-8403-70947f6fec63/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:18 crc kubenswrapper[4492]: I0220 09:23:18.294454 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7c759b9c5-fr6n5_b70a74af-cdf1-4c6a-92ce-1148516ae7ee/dnsmasq-dns/0.log" Feb 20 09:23:18 crc kubenswrapper[4492]: I0220 09:23:18.451839 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0f6bd7f5-2c98-402f-b8ab-c72e291265e3/glance-httpd/0.log" Feb 20 09:23:18 crc kubenswrapper[4492]: I0220 09:23:18.473533 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0f6bd7f5-2c98-402f-b8ab-c72e291265e3/glance-log/0.log" Feb 20 09:23:18 crc kubenswrapper[4492]: I0220 09:23:18.692996 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_377d765a-232c-4fe2-a066-6865ccd8d677/glance-httpd/0.log" Feb 20 09:23:18 crc kubenswrapper[4492]: I0220 09:23:18.734019 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_377d765a-232c-4fe2-a066-6865ccd8d677/glance-log/0.log" Feb 20 09:23:19 crc kubenswrapper[4492]: I0220 09:23:19.826265 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-749c7585d6-zr8k4_618181e6-1990-44c6-b7c5-389d5609252d/heat-engine/0.log" Feb 20 09:23:19 crc kubenswrapper[4492]: I0220 09:23:19.909371 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-d44b9d589-wkjn4_188920e4-7af8-4821-9428-d6224cfeba8e/heat-api/0.log" Feb 20 09:23:20 crc kubenswrapper[4492]: I0220 09:23:20.086943 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d96664ccd-bltcr_22416222-576a-4b0c-a659-145ec65738e3/horizon/0.log" Feb 20 09:23:20 crc kubenswrapper[4492]: I0220 09:23:20.316987 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-s6rgj_3bcfbb69-f489-46a5-b156-1b17dd76fab5/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:20 crc kubenswrapper[4492]: I0220 09:23:20.327132 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-86b9576f68-82n82_270094d8-faff-42ce-9a33-3f9687129320/heat-cfnapi/0.log" Feb 20 09:23:20 crc kubenswrapper[4492]: I0220 09:23:20.560404 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-mdx62_cccd9a33-167d-41ec-b6ab-833d1273f12a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:20 crc kubenswrapper[4492]: I0220 09:23:20.946423 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29526181-rdldl_40796381-0eec-4c7d-b1f0-a24b513ea411/keystone-cron/0.log" Feb 20 09:23:21 crc kubenswrapper[4492]: I0220 09:23:21.065993 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29526241-fgz9q_c8aff97d-0465-4db5-933d-d67cfa494670/keystone-cron/0.log" Feb 20 09:23:21 crc kubenswrapper[4492]: I0220 09:23:21.308101 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d96664ccd-bltcr_22416222-576a-4b0c-a659-145ec65738e3/horizon-log/0.log" Feb 20 09:23:21 crc kubenswrapper[4492]: I0220 09:23:21.495752 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29526301-gkh5l_c66090c8-b2eb-45ae-9504-2d11f6cd1819/keystone-cron/0.log" Feb 20 09:23:21 crc kubenswrapper[4492]: I0220 09:23:21.683004 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_97f89595-4092-4ed0-acc5-421443827a47/kube-state-metrics/0.log" Feb 20 09:23:21 crc kubenswrapper[4492]: I0220 09:23:21.831905 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-bg5w6_94e8423b-3881-4585-8f99-ef80faebae3c/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:21 crc kubenswrapper[4492]: I0220 09:23:21.933121 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7fcbbc74fb-jdvms_4c87d9c3-ee9a-430b-a943-e92ea78e65a5/keystone-api/0.log" Feb 20 09:23:22 crc kubenswrapper[4492]: I0220 09:23:22.251165 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d4db78989-6hrcq_dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166/neutron-httpd/0.log" Feb 20 09:23:22 crc kubenswrapper[4492]: I0220 09:23:22.407187 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-ppnqs_e2680c81-598b-42bf-976e-2c595d48a5b7/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:23 crc kubenswrapper[4492]: I0220 09:23:23.047466 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5d4db78989-6hrcq_dde1c2ae-ef84-46c0-b3e3-fd8aa23c2166/neutron-api/0.log" Feb 20 09:23:23 crc kubenswrapper[4492]: I0220 09:23:23.793437 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8fb65e8c-429a-4279-a277-3730fa1fcd42/nova-cell0-conductor-conductor/0.log" Feb 20 09:23:23 crc kubenswrapper[4492]: I0220 09:23:23.940774 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_eca0914d-a52b-47b6-aeed-4c14979e3afd/nova-cell1-conductor-conductor/0.log" Feb 20 09:23:24 crc kubenswrapper[4492]: I0220 09:23:24.501431 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e720a928-304b-4ff4-9103-40ec152eaf60/nova-cell1-novncproxy-novncproxy/0.log" Feb 20 09:23:24 crc kubenswrapper[4492]: I0220 09:23:24.685819 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-jqcw2_51cf4d1e-7eb2-4882-bd28-c34e75100a67/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:25 crc kubenswrapper[4492]: I0220 09:23:25.075525 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a/nova-metadata-log/0.log" Feb 20 09:23:25 crc kubenswrapper[4492]: I0220 09:23:25.637172 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e9591706-ed64-406e-b33f-d9f59db3c9c7/nova-api-log/0.log" Feb 20 09:23:26 crc kubenswrapper[4492]: I0220 09:23:26.276423 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4f0f24b3-67f0-4232-a358-0e1ae0b25f2f/mysql-bootstrap/0.log" Feb 20 09:23:26 crc kubenswrapper[4492]: I0220 09:23:26.545316 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4f0f24b3-67f0-4232-a358-0e1ae0b25f2f/mysql-bootstrap/0.log" Feb 20 09:23:26 crc kubenswrapper[4492]: I0220 09:23:26.638347 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1c8d2876-8ebc-46af-aa6a-5b03bb8ee207/nova-scheduler-scheduler/0.log" Feb 20 09:23:26 crc kubenswrapper[4492]: I0220 09:23:26.782142 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4f0f24b3-67f0-4232-a358-0e1ae0b25f2f/galera/0.log" Feb 20 09:23:27 crc kubenswrapper[4492]: I0220 09:23:27.121940 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a40f3503-0190-40c1-bbfb-17a454347633/mysql-bootstrap/0.log" Feb 20 09:23:27 crc kubenswrapper[4492]: I0220 09:23:27.281831 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a40f3503-0190-40c1-bbfb-17a454347633/mysql-bootstrap/0.log" Feb 20 09:23:27 crc kubenswrapper[4492]: I0220 09:23:27.340516 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e9591706-ed64-406e-b33f-d9f59db3c9c7/nova-api-api/0.log" Feb 20 09:23:27 crc kubenswrapper[4492]: I0220 09:23:27.764804 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a40f3503-0190-40c1-bbfb-17a454347633/galera/0.log" Feb 20 09:23:27 crc kubenswrapper[4492]: I0220 09:23:27.934273 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_aa29dc13-c6f5-43ef-b0a9-44b6735922d7/openstackclient/0.log" Feb 20 09:23:28 crc kubenswrapper[4492]: I0220 09:23:28.299294 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6k7g9_53c1e7c7-1b2d-4e23-b4bf-fca7fe5833eb/ovn-controller/0.log" Feb 20 09:23:28 crc kubenswrapper[4492]: I0220 09:23:28.461834 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-zwhpv_d4345dc0-8026-48f6-ab0f-57d54ae69433/openstack-network-exporter/0.log" Feb 20 09:23:28 crc kubenswrapper[4492]: I0220 09:23:28.771950 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7csh_3778ef17-0366-4509-8b07-64c301e9de60/ovsdb-server-init/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.014875 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7csh_3778ef17-0366-4509-8b07-64c301e9de60/ovs-vswitchd/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.054615 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7csh_3778ef17-0366-4509-8b07-64c301e9de60/ovsdb-server/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.055299 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t7csh_3778ef17-0366-4509-8b07-64c301e9de60/ovsdb-server-init/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.395988 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-75vqr_1f5d5f14-0d7b-4117-9651-c1b16f9722f1/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.710555 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_28bcc602-c9ec-4bfe-aed6-b45f16a327db/openstack-network-exporter/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.713751 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_28bcc602-c9ec-4bfe-aed6-b45f16a327db/ovn-northd/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.964249 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3b8d9698-66bb-44b6-b225-e77fc088ff02/ovsdbserver-nb/0.log" Feb 20 09:23:29 crc kubenswrapper[4492]: I0220 09:23:29.996677 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3b8d9698-66bb-44b6-b225-e77fc088ff02/openstack-network-exporter/0.log" Feb 20 09:23:30 crc kubenswrapper[4492]: I0220 09:23:30.255290 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbdd3644-632d-4bf1-90d0-0fb9a3392884/openstack-network-exporter/0.log" Feb 20 09:23:30 crc kubenswrapper[4492]: I0220 09:23:30.284775 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_dbdd3644-632d-4bf1-90d0-0fb9a3392884/ovsdbserver-sb/0.log" Feb 20 09:23:30 crc kubenswrapper[4492]: I0220 09:23:30.557562 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_47f7dfdc-8c6b-4dcb-83e0-0a1ee209f03a/nova-metadata-metadata/0.log" Feb 20 09:23:30 crc kubenswrapper[4492]: I0220 09:23:30.899747 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_18f922e8-15c1-49a7-af1b-38a85b924404/setup-container/0.log" Feb 20 09:23:31 crc kubenswrapper[4492]: I0220 09:23:31.085606 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_18f922e8-15c1-49a7-af1b-38a85b924404/setup-container/0.log" Feb 20 09:23:31 crc kubenswrapper[4492]: I0220 09:23:31.215224 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_18f922e8-15c1-49a7-af1b-38a85b924404/rabbitmq/0.log" Feb 20 09:23:31 crc kubenswrapper[4492]: I0220 09:23:31.248886 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65bf8dfc4b-nwxp9_eacd0a76-6ee4-42ea-b2b8-7faef8c27f68/placement-api/0.log" Feb 20 09:23:31 crc kubenswrapper[4492]: I0220 09:23:31.370931 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65bf8dfc4b-nwxp9_eacd0a76-6ee4-42ea-b2b8-7faef8c27f68/placement-log/0.log" Feb 20 09:23:31 crc kubenswrapper[4492]: I0220 09:23:31.818363 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a4dfa138-8cc8-40e0-b2df-21449a994146/setup-container/0.log" Feb 20 09:23:32 crc kubenswrapper[4492]: I0220 09:23:32.133405 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a4dfa138-8cc8-40e0-b2df-21449a994146/setup-container/0.log" Feb 20 09:23:32 crc kubenswrapper[4492]: I0220 09:23:32.154969 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a4dfa138-8cc8-40e0-b2df-21449a994146/rabbitmq/0.log" Feb 20 09:23:32 crc kubenswrapper[4492]: I0220 09:23:32.166976 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-l565w_b68494be-4b9d-4931-ad0a-5899e5ccf709/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:32 crc kubenswrapper[4492]: I0220 09:23:32.449833 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qgbrg_d9bdd400-95f5-40ca-b1d3-f1f2ddcbd793/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:32 crc kubenswrapper[4492]: I0220 09:23:32.486393 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jwpxd_a54176de-a2a7-4819-aa62-f5da167b49c0/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:32 crc kubenswrapper[4492]: I0220 09:23:32.759643 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-92sh4_2505a4cb-335b-4570-9d5a-f0af6666ea9b/ssh-known-hosts-edpm-deployment/0.log" Feb 20 09:23:32 crc kubenswrapper[4492]: I0220 09:23:32.769587 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-g2clh_193682fc-046c-4816-a3cf-610626b53e6e/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:33 crc kubenswrapper[4492]: I0220 09:23:33.107414 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c8d654db7-jfjgs_b14dc392-054d-4257-b340-642e7527aa67/proxy-server/0.log" Feb 20 09:23:33 crc kubenswrapper[4492]: I0220 09:23:33.493794 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7c8d654db7-jfjgs_b14dc392-054d-4257-b340-642e7527aa67/proxy-httpd/0.log" Feb 20 09:23:33 crc kubenswrapper[4492]: I0220 09:23:33.885733 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-sftjr_b39c38ec-fe4b-407f-b4d8-b5c60049b4b5/swift-ring-rebalance/0.log" Feb 20 09:23:33 crc kubenswrapper[4492]: I0220 09:23:33.945154 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/account-reaper/0.log" Feb 20 09:23:33 crc kubenswrapper[4492]: I0220 09:23:33.979919 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/account-auditor/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.212226 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/account-server/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.298549 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/container-auditor/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.338380 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/account-replicator/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.427553 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/container-replicator/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.470406 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/container-server/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.608999 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/container-updater/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.710452 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/object-auditor/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.864109 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/object-expirer/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.887997 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/object-replicator/0.log" Feb 20 09:23:34 crc kubenswrapper[4492]: I0220 09:23:34.982042 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/object-server/0.log" Feb 20 09:23:35 crc kubenswrapper[4492]: I0220 09:23:35.059593 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/object-updater/0.log" Feb 20 09:23:35 crc kubenswrapper[4492]: I0220 09:23:35.370455 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/rsync/0.log" Feb 20 09:23:35 crc kubenswrapper[4492]: I0220 09:23:35.664019 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_12f20cf1-c22c-48a2-9d5f-45f02009d0a5/swift-recon-cron/0.log" Feb 20 09:23:35 crc kubenswrapper[4492]: I0220 09:23:35.829431 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-npbqt_a79e5d8f-5072-4056-b1e3-b7853f563978/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:36 crc kubenswrapper[4492]: I0220 09:23:36.073710 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-multi-thread-testing_419ec593-81d8-44f8-9d19-5df9ebf183e1/tempest-tests-tempest-tests-runner/0.log" Feb 20 09:23:36 crc kubenswrapper[4492]: I0220 09:23:36.158041 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-thread-testing_0d1fd559-3c53-44c1-aa9b-556b306a479b/tempest-tests-tempest-tests-runner/0.log" Feb 20 09:23:36 crc kubenswrapper[4492]: I0220 09:23:36.323129 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_2ae49fdb-33f1-44a8-a26f-222c204cdfac/test-operator-logs-container/0.log" Feb 20 09:23:36 crc kubenswrapper[4492]: I0220 09:23:36.466880 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-9jbjg_62a602f6-1225-4dc5-9959-8521e0d07631/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 20 09:23:39 crc kubenswrapper[4492]: I0220 09:23:39.310912 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:23:39 crc kubenswrapper[4492]: I0220 09:23:39.311255 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:23:39 crc kubenswrapper[4492]: I0220 09:23:39.311324 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 09:23:39 crc kubenswrapper[4492]: I0220 09:23:39.312036 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:23:39 crc kubenswrapper[4492]: I0220 09:23:39.312098 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" gracePeriod=600 Feb 20 09:23:39 crc kubenswrapper[4492]: E0220 09:23:39.488425 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:23:40 crc kubenswrapper[4492]: I0220 09:23:40.171906 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" exitCode=0 Feb 20 09:23:40 crc kubenswrapper[4492]: I0220 09:23:40.172237 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a"} Feb 20 09:23:40 crc kubenswrapper[4492]: I0220 09:23:40.172702 4492 scope.go:117] "RemoveContainer" containerID="6ff6c52a7e322eb6e1d90878291c8341550e11afc0ae66488e8cec367610b76a" Feb 20 09:23:40 crc kubenswrapper[4492]: I0220 09:23:40.173970 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:23:40 crc kubenswrapper[4492]: E0220 09:23:40.174818 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:23:48 crc kubenswrapper[4492]: I0220 09:23:48.772820 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_522bd757-e09f-424a-8459-57a7fb68952f/memcached/0.log" Feb 20 09:23:53 crc kubenswrapper[4492]: I0220 09:23:53.563490 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:23:53 crc kubenswrapper[4492]: E0220 09:23:53.564381 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:24:05 crc kubenswrapper[4492]: I0220 09:24:05.557528 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:24:05 crc kubenswrapper[4492]: E0220 09:24:05.558606 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:24:10 crc kubenswrapper[4492]: I0220 09:24:10.274291 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6_dc079aa5-a7a5-4043-951b-65da547fa1e1/util/0.log" Feb 20 09:24:10 crc kubenswrapper[4492]: I0220 09:24:10.486503 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6_dc079aa5-a7a5-4043-951b-65da547fa1e1/pull/0.log" Feb 20 09:24:10 crc kubenswrapper[4492]: I0220 09:24:10.490129 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6_dc079aa5-a7a5-4043-951b-65da547fa1e1/util/0.log" Feb 20 09:24:10 crc kubenswrapper[4492]: I0220 09:24:10.624586 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6_dc079aa5-a7a5-4043-951b-65da547fa1e1/pull/0.log" Feb 20 09:24:10 crc kubenswrapper[4492]: I0220 09:24:10.749384 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6_dc079aa5-a7a5-4043-951b-65da547fa1e1/util/0.log" Feb 20 09:24:10 crc kubenswrapper[4492]: I0220 09:24:10.775754 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6_dc079aa5-a7a5-4043-951b-65da547fa1e1/pull/0.log" Feb 20 09:24:10 crc kubenswrapper[4492]: I0220 09:24:10.791079 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8f52c407bdc9ecc5c9ed04cde121370cff57ca187d042afc6ea79b7967hkhg6_dc079aa5-a7a5-4043-951b-65da547fa1e1/extract/0.log" Feb 20 09:24:11 crc kubenswrapper[4492]: I0220 09:24:11.936971 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-h27fh_16dae257-55aa-40f7-9308-182cd7ff886b/manager/0.log" Feb 20 09:24:12 crc kubenswrapper[4492]: I0220 09:24:12.487250 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987464f4-f5lt2_dd787991-93c0-4baf-8a53-c2ec22c09ede/manager/0.log" Feb 20 09:24:12 crc kubenswrapper[4492]: I0220 09:24:12.873352 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-5kqdb_80579e11-ccf9-4377-a013-345a55ac33ab/manager/0.log" Feb 20 09:24:13 crc kubenswrapper[4492]: I0220 09:24:13.200183 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-jm5rz_9c26743a-8dbb-4bdc-a2a8-430efd8070c6/manager/0.log" Feb 20 09:24:13 crc kubenswrapper[4492]: I0220 09:24:13.845618 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-s7bxs_86409721-594f-4c20-9f11-2f89572c3036/manager/0.log" Feb 20 09:24:14 crc kubenswrapper[4492]: I0220 09:24:14.011901 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-szmrb_bab88145-f203-4b8c-b9d3-a03ee1af5ac7/manager/0.log" Feb 20 09:24:14 crc kubenswrapper[4492]: I0220 09:24:14.295859 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-t8cd6_2918bb03-8106-4586-90bc-2618b3e78239/manager/0.log" Feb 20 09:24:14 crc kubenswrapper[4492]: I0220 09:24:14.493653 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-54f6768c69-dntvf_7e6bf429-fb74-4a04-ac01-b4f60d4c00c2/manager/0.log" Feb 20 09:24:14 crc kubenswrapper[4492]: I0220 09:24:14.692107 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-dnt2c_b1cfd35f-eae6-453c-8687-6bf96945edf2/manager/0.log" Feb 20 09:24:14 crc kubenswrapper[4492]: I0220 09:24:14.955821 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-64ddbf8bb-vckjp_ea530a5e-2861-4559-b80a-ef8589d9788f/manager/0.log" Feb 20 09:24:15 crc kubenswrapper[4492]: I0220 09:24:15.136064 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-2zzhs_883be441-1faf-4b78-9c22-16c19986902a/manager/0.log" Feb 20 09:24:15 crc kubenswrapper[4492]: I0220 09:24:15.566172 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-5d946d989d-vqxgx_1f1dd67b-1f84-49a3-8ffa-bba932485c90/manager/0.log" Feb 20 09:24:15 crc kubenswrapper[4492]: I0220 09:24:15.606429 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-fb5fcc5b8-l5vdv_41543a2e-55d9-465c-bb77-f2d7b04b7e2b/manager/0.log" Feb 20 09:24:15 crc kubenswrapper[4492]: I0220 09:24:15.977613 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6679bf9b57-hd7zh_c7b6f955-806d-4f15-bc40-8fccbdeefa08/operator/0.log" Feb 20 09:24:16 crc kubenswrapper[4492]: I0220 09:24:16.647512 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4hv4z_7fd5d9fb-a07c-4715-b281-d181fd8f5543/registry-server/0.log" Feb 20 09:24:16 crc kubenswrapper[4492]: I0220 09:24:16.986852 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-d44cf6b75-t5t76_91e88c5d-7cb2-4ebd-9a96-d3620ea38c97/manager/0.log" Feb 20 09:24:17 crc kubenswrapper[4492]: I0220 09:24:17.280352 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-6hbwt_10788676-0981-46c0-8a6a-eda4f457e117/manager/0.log" Feb 20 09:24:17 crc kubenswrapper[4492]: I0220 09:24:17.567152 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-lk9pm_a67d76ca-f5c0-4416-a85b-9de9211aed83/operator/0.log" Feb 20 09:24:17 crc kubenswrapper[4492]: I0220 09:24:17.838137 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-77tpg_ba9af5af-1c2b-421c-a59a-4f4a203a8e3e/manager/0.log" Feb 20 09:24:18 crc kubenswrapper[4492]: I0220 09:24:18.166551 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-7f45b4ff68-t4h7v_e2b03896-ea8a-469a-8a43-4f04f2108071/manager/0.log" Feb 20 09:24:18 crc kubenswrapper[4492]: I0220 09:24:18.444288 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-7866795846-p5tgj_c324ea9b-913d-4766-ac5a-790941b21763/manager/0.log" Feb 20 09:24:18 crc kubenswrapper[4492]: I0220 09:24:18.751940 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5db88f68c-4h4j7_7b3527ea-31e6-4b64-b5b1-1e683fe05689/manager/0.log" Feb 20 09:24:18 crc kubenswrapper[4492]: I0220 09:24:18.816578 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-69ff7bc449-242g8_3bd31619-5d0d-4da6-bc6d-3f7fab1a67cf/manager/0.log" Feb 20 09:24:19 crc kubenswrapper[4492]: I0220 09:24:19.562820 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:24:19 crc kubenswrapper[4492]: E0220 09:24:19.563165 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:24:19 crc kubenswrapper[4492]: I0220 09:24:19.754373 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f8888797-f6zqc_cafb9e27-7f46-4d0d-a27f-7f60d11d5b79/manager/0.log" Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.827128 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4q8dt"] Feb 20 09:24:25 crc kubenswrapper[4492]: E0220 09:24:25.828493 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40b6c08e-ba49-4b3d-a24e-aadf94ba8244" containerName="container-00" Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.828512 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="40b6c08e-ba49-4b3d-a24e-aadf94ba8244" containerName="container-00" Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.828744 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="40b6c08e-ba49-4b3d-a24e-aadf94ba8244" containerName="container-00" Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.832489 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.876395 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q8dt"] Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.943696 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dxxr\" (UniqueName: \"kubernetes.io/projected/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-kube-api-access-2dxxr\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.943774 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-utilities\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:25 crc kubenswrapper[4492]: I0220 09:24:25.943900 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-catalog-content\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:26 crc kubenswrapper[4492]: I0220 09:24:26.046455 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dxxr\" (UniqueName: \"kubernetes.io/projected/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-kube-api-access-2dxxr\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:26 crc kubenswrapper[4492]: I0220 09:24:26.046584 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-utilities\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:26 crc kubenswrapper[4492]: I0220 09:24:26.046883 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-catalog-content\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:26 crc kubenswrapper[4492]: I0220 09:24:26.047412 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-catalog-content\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:26 crc kubenswrapper[4492]: I0220 09:24:26.048067 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-utilities\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:26 crc kubenswrapper[4492]: I0220 09:24:26.086336 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dxxr\" (UniqueName: \"kubernetes.io/projected/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-kube-api-access-2dxxr\") pod \"redhat-marketplace-4q8dt\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:26 crc kubenswrapper[4492]: I0220 09:24:26.166404 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:27 crc kubenswrapper[4492]: I0220 09:24:27.052493 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q8dt"] Feb 20 09:24:27 crc kubenswrapper[4492]: E0220 09:24:27.371457 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf6f27b3_1de0_4da3_b1f7_070c36eba0e1.slice/crio-conmon-8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d.scope\": RecentStats: unable to find data in memory cache]" Feb 20 09:24:27 crc kubenswrapper[4492]: I0220 09:24:27.571595 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-qrmfx_d1f3b23e-c05b-42f0-af00-5af28a768c0e/manager/0.log" Feb 20 09:24:27 crc kubenswrapper[4492]: I0220 09:24:27.676672 4492 generic.go:334] "Generic (PLEG): container finished" podID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerID="8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d" exitCode=0 Feb 20 09:24:27 crc kubenswrapper[4492]: I0220 09:24:27.676949 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q8dt" event={"ID":"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1","Type":"ContainerDied","Data":"8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d"} Feb 20 09:24:27 crc kubenswrapper[4492]: I0220 09:24:27.676986 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q8dt" event={"ID":"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1","Type":"ContainerStarted","Data":"e4005292ffb7ec8cc50b6fe86d2e4b0c39f85cc568fc196273f0e2cddba760da"} Feb 20 09:24:28 crc kubenswrapper[4492]: I0220 09:24:28.690375 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q8dt" event={"ID":"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1","Type":"ContainerStarted","Data":"3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917"} Feb 20 09:24:29 crc kubenswrapper[4492]: I0220 09:24:29.701299 4492 generic.go:334] "Generic (PLEG): container finished" podID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerID="3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917" exitCode=0 Feb 20 09:24:29 crc kubenswrapper[4492]: I0220 09:24:29.701402 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q8dt" event={"ID":"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1","Type":"ContainerDied","Data":"3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917"} Feb 20 09:24:30 crc kubenswrapper[4492]: I0220 09:24:30.713264 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q8dt" event={"ID":"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1","Type":"ContainerStarted","Data":"855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550"} Feb 20 09:24:30 crc kubenswrapper[4492]: I0220 09:24:30.732201 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4q8dt" podStartSLOduration=3.221384856 podStartE2EDuration="5.732185686s" podCreationTimestamp="2026-02-20 09:24:25 +0000 UTC" firstStartedPulling="2026-02-20 09:24:27.680131003 +0000 UTC m=+9824.451419982" lastFinishedPulling="2026-02-20 09:24:30.190931834 +0000 UTC m=+9826.962220812" observedRunningTime="2026-02-20 09:24:30.72960522 +0000 UTC m=+9827.500894198" watchObservedRunningTime="2026-02-20 09:24:30.732185686 +0000 UTC m=+9827.503474664" Feb 20 09:24:34 crc kubenswrapper[4492]: I0220 09:24:34.557513 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:24:34 crc kubenswrapper[4492]: E0220 09:24:34.559574 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:24:36 crc kubenswrapper[4492]: I0220 09:24:36.167199 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:36 crc kubenswrapper[4492]: I0220 09:24:36.167669 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:36 crc kubenswrapper[4492]: I0220 09:24:36.210082 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:36 crc kubenswrapper[4492]: I0220 09:24:36.811459 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:36 crc kubenswrapper[4492]: I0220 09:24:36.871881 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q8dt"] Feb 20 09:24:38 crc kubenswrapper[4492]: I0220 09:24:38.785957 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4q8dt" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="registry-server" containerID="cri-o://855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550" gracePeriod=2 Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.248997 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.268713 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-catalog-content\") pod \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.268963 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dxxr\" (UniqueName: \"kubernetes.io/projected/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-kube-api-access-2dxxr\") pod \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.269054 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-utilities\") pod \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\" (UID: \"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1\") " Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.269422 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-utilities" (OuterVolumeSpecName: "utilities") pod "bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" (UID: "bf6f27b3-1de0-4da3-b1f7-070c36eba0e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.270033 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.277058 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-kube-api-access-2dxxr" (OuterVolumeSpecName: "kube-api-access-2dxxr") pod "bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" (UID: "bf6f27b3-1de0-4da3-b1f7-070c36eba0e1"). InnerVolumeSpecName "kube-api-access-2dxxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.285199 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" (UID: "bf6f27b3-1de0-4da3-b1f7-070c36eba0e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.373134 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dxxr\" (UniqueName: \"kubernetes.io/projected/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-kube-api-access-2dxxr\") on node \"crc\" DevicePath \"\"" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.373175 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.794988 4492 generic.go:334] "Generic (PLEG): container finished" podID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerID="855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550" exitCode=0 Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.795294 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q8dt" event={"ID":"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1","Type":"ContainerDied","Data":"855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550"} Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.795336 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4q8dt" event={"ID":"bf6f27b3-1de0-4da3-b1f7-070c36eba0e1","Type":"ContainerDied","Data":"e4005292ffb7ec8cc50b6fe86d2e4b0c39f85cc568fc196273f0e2cddba760da"} Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.795354 4492 scope.go:117] "RemoveContainer" containerID="855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.795490 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4q8dt" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.828842 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q8dt"] Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.829228 4492 scope.go:117] "RemoveContainer" containerID="3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.850391 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4q8dt"] Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.861499 4492 scope.go:117] "RemoveContainer" containerID="8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.893156 4492 scope.go:117] "RemoveContainer" containerID="855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550" Feb 20 09:24:39 crc kubenswrapper[4492]: E0220 09:24:39.894452 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550\": container with ID starting with 855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550 not found: ID does not exist" containerID="855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.894503 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550"} err="failed to get container status \"855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550\": rpc error: code = NotFound desc = could not find container \"855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550\": container with ID starting with 855c8b7fd9f6fd1979fef46ca161a081e9574c26933b12b519e5c1a6f76fa550 not found: ID does not exist" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.894525 4492 scope.go:117] "RemoveContainer" containerID="3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917" Feb 20 09:24:39 crc kubenswrapper[4492]: E0220 09:24:39.894818 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917\": container with ID starting with 3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917 not found: ID does not exist" containerID="3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.894860 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917"} err="failed to get container status \"3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917\": rpc error: code = NotFound desc = could not find container \"3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917\": container with ID starting with 3f35c06c9794e4ffc0e26c4f911f66bac35737270eb7d1861bf6f8ec949b9917 not found: ID does not exist" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.894891 4492 scope.go:117] "RemoveContainer" containerID="8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d" Feb 20 09:24:39 crc kubenswrapper[4492]: E0220 09:24:39.895174 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d\": container with ID starting with 8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d not found: ID does not exist" containerID="8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d" Feb 20 09:24:39 crc kubenswrapper[4492]: I0220 09:24:39.895199 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d"} err="failed to get container status \"8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d\": rpc error: code = NotFound desc = could not find container \"8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d\": container with ID starting with 8608fcab935e9187bb7dfe9c87cae48840364287b1318875a739e8c6edfc303d not found: ID does not exist" Feb 20 09:24:41 crc kubenswrapper[4492]: I0220 09:24:41.567195 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" path="/var/lib/kubelet/pods/bf6f27b3-1de0-4da3-b1f7-070c36eba0e1/volumes" Feb 20 09:24:44 crc kubenswrapper[4492]: I0220 09:24:44.046690 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zc5kp_a2bc1973-e22e-47e6-88bf-edab0829592e/control-plane-machine-set-operator/0.log" Feb 20 09:24:44 crc kubenswrapper[4492]: I0220 09:24:44.242708 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sjzvg_4c3dd185-faf8-4150-81b6-ef352c48148b/machine-api-operator/0.log" Feb 20 09:24:44 crc kubenswrapper[4492]: I0220 09:24:44.276375 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sjzvg_4c3dd185-faf8-4150-81b6-ef352c48148b/kube-rbac-proxy/0.log" Feb 20 09:24:48 crc kubenswrapper[4492]: I0220 09:24:48.557308 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:24:48 crc kubenswrapper[4492]: E0220 09:24:48.557893 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:24:58 crc kubenswrapper[4492]: I0220 09:24:58.115536 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-fgbwd_63b72bed-a101-41fc-8402-f12eb5a6420b/cert-manager-controller/0.log" Feb 20 09:24:58 crc kubenswrapper[4492]: I0220 09:24:58.355869 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-68zkr_3e8ca167-0346-42b9-8e8e-48b1c85b3f5a/cert-manager-cainjector/0.log" Feb 20 09:24:58 crc kubenswrapper[4492]: I0220 09:24:58.432990 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-l7wk4_b43a3a53-5167-477a-9531-ce42d03c5494/cert-manager-webhook/0.log" Feb 20 09:25:00 crc kubenswrapper[4492]: I0220 09:25:00.557191 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:25:00 crc kubenswrapper[4492]: E0220 09:25:00.558008 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:25:11 crc kubenswrapper[4492]: I0220 09:25:11.836205 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5c78fc5d65-2vhlz_69627a87-2af9-40ee-860b-52cc31c8591d/nmstate-console-plugin/0.log" Feb 20 09:25:12 crc kubenswrapper[4492]: I0220 09:25:12.042165 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-948rj_cf9c83fb-58e9-4c00-9c3b-9fcaa4da2914/nmstate-handler/0.log" Feb 20 09:25:12 crc kubenswrapper[4492]: I0220 09:25:12.124121 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-j8ddx_0378fe77-c8df-4601-88af-0407d45e8d13/kube-rbac-proxy/0.log" Feb 20 09:25:12 crc kubenswrapper[4492]: I0220 09:25:12.183613 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58c85c668d-j8ddx_0378fe77-c8df-4601-88af-0407d45e8d13/nmstate-metrics/0.log" Feb 20 09:25:12 crc kubenswrapper[4492]: I0220 09:25:12.376656 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-694c9596b7-4btwm_026317d3-c00d-420e-b6cc-f5beba8f1e7f/nmstate-operator/0.log" Feb 20 09:25:12 crc kubenswrapper[4492]: I0220 09:25:12.442285 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-866bcb46dc-tp9zd_07c86583-b824-4484-9ddf-d4614fa1a99c/nmstate-webhook/0.log" Feb 20 09:25:14 crc kubenswrapper[4492]: I0220 09:25:14.556695 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:25:14 crc kubenswrapper[4492]: E0220 09:25:14.577801 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:25:29 crc kubenswrapper[4492]: I0220 09:25:29.560216 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:25:29 crc kubenswrapper[4492]: E0220 09:25:29.561692 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:25:40 crc kubenswrapper[4492]: I0220 09:25:40.587301 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-28vtj_6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3/kube-rbac-proxy/0.log" Feb 20 09:25:40 crc kubenswrapper[4492]: I0220 09:25:40.706632 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-28vtj_6cf02e0b-a76b-491d-9e7e-fa147c7f0aa3/controller/0.log" Feb 20 09:25:40 crc kubenswrapper[4492]: I0220 09:25:40.807446 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-frr-files/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.016363 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-frr-files/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.052198 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-reloader/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.053497 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-reloader/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.104271 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-metrics/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.288721 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-reloader/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.306886 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-frr-files/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.367245 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-metrics/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.383469 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-metrics/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.579740 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-reloader/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.587584 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-metrics/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.592332 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/cp-frr-files/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.633780 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/controller/0.log" Feb 20 09:25:41 crc kubenswrapper[4492]: I0220 09:25:41.765235 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/frr-metrics/0.log" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.008504 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/kube-rbac-proxy-frr/0.log" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.034271 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/reloader/0.log" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.045585 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/kube-rbac-proxy/0.log" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.316360 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-lmpxr_c3b6f338-04d7-40c8-baa6-8f17a228a458/frr-k8s-webhook-server/0.log" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.556594 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:25:42 crc kubenswrapper[4492]: E0220 09:25:42.556928 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.631297 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-76ffddff97-9dv5b_2fe223a3-3ace-4d01-8379-d74032c024a0/manager/0.log" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.781892 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7d989b6fcf-r2r6v_9ef7522d-21b4-4acf-9cf3-c0217fcc173b/webhook-server/0.log" Feb 20 09:25:42 crc kubenswrapper[4492]: I0220 09:25:42.974753 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rnqvv_7837488e-131f-4866-9417-5c9605902130/kube-rbac-proxy/0.log" Feb 20 09:25:43 crc kubenswrapper[4492]: I0220 09:25:43.852092 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rnqvv_7837488e-131f-4866-9417-5c9605902130/speaker/0.log" Feb 20 09:25:44 crc kubenswrapper[4492]: I0220 09:25:44.003687 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sps8s_1224d7af-027f-4a22-b675-606a9650f716/frr/0.log" Feb 20 09:25:53 crc kubenswrapper[4492]: I0220 09:25:53.561948 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:25:53 crc kubenswrapper[4492]: E0220 09:25:53.563145 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.129533 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q_d4e34ff6-7469-44df-8ab0-fdb3f6d0535c/util/0.log" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.462738 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q_d4e34ff6-7469-44df-8ab0-fdb3f6d0535c/util/0.log" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.464095 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q_d4e34ff6-7469-44df-8ab0-fdb3f6d0535c/pull/0.log" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.501446 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q_d4e34ff6-7469-44df-8ab0-fdb3f6d0535c/pull/0.log" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.731802 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q_d4e34ff6-7469-44df-8ab0-fdb3f6d0535c/util/0.log" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.765824 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q_d4e34ff6-7469-44df-8ab0-fdb3f6d0535c/pull/0.log" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.788244 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc2135v28q_d4e34ff6-7469-44df-8ab0-fdb3f6d0535c/extract/0.log" Feb 20 09:25:59 crc kubenswrapper[4492]: I0220 09:25:59.924307 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vc5ws_b0e828a1-a902-411f-8b55-f12b3a8c3e1c/extract-utilities/0.log" Feb 20 09:26:00 crc kubenswrapper[4492]: I0220 09:26:00.282385 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vc5ws_b0e828a1-a902-411f-8b55-f12b3a8c3e1c/extract-content/0.log" Feb 20 09:26:00 crc kubenswrapper[4492]: I0220 09:26:00.317592 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vc5ws_b0e828a1-a902-411f-8b55-f12b3a8c3e1c/extract-content/0.log" Feb 20 09:26:00 crc kubenswrapper[4492]: I0220 09:26:00.373248 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vc5ws_b0e828a1-a902-411f-8b55-f12b3a8c3e1c/extract-utilities/0.log" Feb 20 09:26:00 crc kubenswrapper[4492]: I0220 09:26:00.638612 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vc5ws_b0e828a1-a902-411f-8b55-f12b3a8c3e1c/extract-content/0.log" Feb 20 09:26:00 crc kubenswrapper[4492]: I0220 09:26:00.678883 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vc5ws_b0e828a1-a902-411f-8b55-f12b3a8c3e1c/extract-utilities/0.log" Feb 20 09:26:00 crc kubenswrapper[4492]: I0220 09:26:00.870067 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-67rg8_922456bc-c937-4182-b66c-9b6cc71f80a4/extract-utilities/0.log" Feb 20 09:26:01 crc kubenswrapper[4492]: I0220 09:26:01.165674 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-67rg8_922456bc-c937-4182-b66c-9b6cc71f80a4/extract-utilities/0.log" Feb 20 09:26:01 crc kubenswrapper[4492]: I0220 09:26:01.347401 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-67rg8_922456bc-c937-4182-b66c-9b6cc71f80a4/extract-content/0.log" Feb 20 09:26:01 crc kubenswrapper[4492]: I0220 09:26:01.398203 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-67rg8_922456bc-c937-4182-b66c-9b6cc71f80a4/extract-content/0.log" Feb 20 09:26:01 crc kubenswrapper[4492]: I0220 09:26:01.598214 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-67rg8_922456bc-c937-4182-b66c-9b6cc71f80a4/extract-utilities/0.log" Feb 20 09:26:01 crc kubenswrapper[4492]: I0220 09:26:01.655164 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-67rg8_922456bc-c937-4182-b66c-9b6cc71f80a4/extract-content/0.log" Feb 20 09:26:01 crc kubenswrapper[4492]: I0220 09:26:01.691213 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vc5ws_b0e828a1-a902-411f-8b55-f12b3a8c3e1c/registry-server/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.073069 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg_634446d6-13a4-4a8d-9e28-b2abebca10a3/util/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.338654 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-67rg8_922456bc-c937-4182-b66c-9b6cc71f80a4/registry-server/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.418725 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg_634446d6-13a4-4a8d-9e28-b2abebca10a3/pull/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.418817 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg_634446d6-13a4-4a8d-9e28-b2abebca10a3/util/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.458780 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg_634446d6-13a4-4a8d-9e28-b2abebca10a3/pull/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.603930 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg_634446d6-13a4-4a8d-9e28-b2abebca10a3/extract/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.614934 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg_634446d6-13a4-4a8d-9e28-b2abebca10a3/pull/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.618146 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f938df2ce267491f058ea7e3036e97ee3f65bf3665185b1a4f52323ecazgcrg_634446d6-13a4-4a8d-9e28-b2abebca10a3/util/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.885349 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ddlzq_1be6eb66-d6a5-41a8-acf9-8297fefe0279/marketplace-operator/0.log" Feb 20 09:26:02 crc kubenswrapper[4492]: I0220 09:26:02.937337 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cms44_2bcd6b9a-5837-48eb-98d0-691541bfff4d/extract-utilities/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.126614 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cms44_2bcd6b9a-5837-48eb-98d0-691541bfff4d/extract-content/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.146962 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cms44_2bcd6b9a-5837-48eb-98d0-691541bfff4d/extract-utilities/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.188742 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cms44_2bcd6b9a-5837-48eb-98d0-691541bfff4d/extract-content/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.361994 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cms44_2bcd6b9a-5837-48eb-98d0-691541bfff4d/extract-utilities/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.380071 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cms44_2bcd6b9a-5837-48eb-98d0-691541bfff4d/extract-content/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.631144 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5kkhh_9d2d17f0-7864-41dd-8b9f-285944a57376/extract-utilities/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.745871 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-cms44_2bcd6b9a-5837-48eb-98d0-691541bfff4d/registry-server/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.822393 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5kkhh_9d2d17f0-7864-41dd-8b9f-285944a57376/extract-utilities/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.826727 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5kkhh_9d2d17f0-7864-41dd-8b9f-285944a57376/extract-content/0.log" Feb 20 09:26:03 crc kubenswrapper[4492]: I0220 09:26:03.857548 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5kkhh_9d2d17f0-7864-41dd-8b9f-285944a57376/extract-content/0.log" Feb 20 09:26:04 crc kubenswrapper[4492]: I0220 09:26:04.016509 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5kkhh_9d2d17f0-7864-41dd-8b9f-285944a57376/extract-utilities/0.log" Feb 20 09:26:04 crc kubenswrapper[4492]: I0220 09:26:04.031992 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5kkhh_9d2d17f0-7864-41dd-8b9f-285944a57376/extract-content/0.log" Feb 20 09:26:05 crc kubenswrapper[4492]: I0220 09:26:05.053805 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-5kkhh_9d2d17f0-7864-41dd-8b9f-285944a57376/registry-server/0.log" Feb 20 09:26:08 crc kubenswrapper[4492]: I0220 09:26:08.557256 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:26:08 crc kubenswrapper[4492]: E0220 09:26:08.558348 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:26:22 crc kubenswrapper[4492]: I0220 09:26:22.557931 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:26:22 crc kubenswrapper[4492]: E0220 09:26:22.558646 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:26:36 crc kubenswrapper[4492]: I0220 09:26:36.557174 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:26:36 crc kubenswrapper[4492]: E0220 09:26:36.558131 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:26:49 crc kubenswrapper[4492]: I0220 09:26:49.557227 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:26:49 crc kubenswrapper[4492]: E0220 09:26:49.558105 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.492319 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bltlg"] Feb 20 09:27:00 crc kubenswrapper[4492]: E0220 09:27:00.493033 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="extract-utilities" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.493049 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="extract-utilities" Feb 20 09:27:00 crc kubenswrapper[4492]: E0220 09:27:00.493064 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="registry-server" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.493069 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="registry-server" Feb 20 09:27:00 crc kubenswrapper[4492]: E0220 09:27:00.493090 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="extract-content" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.493096 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="extract-content" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.493319 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf6f27b3-1de0-4da3-b1f7-070c36eba0e1" containerName="registry-server" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.495549 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.506341 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bltlg"] Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.517971 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7x7c\" (UniqueName: \"kubernetes.io/projected/d5ee0a49-193d-478a-af87-42c6fb7ea091-kube-api-access-g7x7c\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.518167 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-utilities\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.518544 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-catalog-content\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.622242 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-utilities\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.622521 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-catalog-content\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.622681 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7x7c\" (UniqueName: \"kubernetes.io/projected/d5ee0a49-193d-478a-af87-42c6fb7ea091-kube-api-access-g7x7c\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.623321 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-utilities\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.623405 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-catalog-content\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.654796 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7x7c\" (UniqueName: \"kubernetes.io/projected/d5ee0a49-193d-478a-af87-42c6fb7ea091-kube-api-access-g7x7c\") pod \"redhat-operators-bltlg\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:00 crc kubenswrapper[4492]: I0220 09:27:00.829675 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:01 crc kubenswrapper[4492]: I0220 09:27:01.464821 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bltlg"] Feb 20 09:27:01 crc kubenswrapper[4492]: W0220 09:27:01.481021 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5ee0a49_193d_478a_af87_42c6fb7ea091.slice/crio-7e186b4c85e9255c76cf809801372380f84684432210be646644751017745c85 WatchSource:0}: Error finding container 7e186b4c85e9255c76cf809801372380f84684432210be646644751017745c85: Status 404 returned error can't find the container with id 7e186b4c85e9255c76cf809801372380f84684432210be646644751017745c85 Feb 20 09:27:02 crc kubenswrapper[4492]: I0220 09:27:02.166983 4492 generic.go:334] "Generic (PLEG): container finished" podID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerID="f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80" exitCode=0 Feb 20 09:27:02 crc kubenswrapper[4492]: I0220 09:27:02.167282 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bltlg" event={"ID":"d5ee0a49-193d-478a-af87-42c6fb7ea091","Type":"ContainerDied","Data":"f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80"} Feb 20 09:27:02 crc kubenswrapper[4492]: I0220 09:27:02.167318 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bltlg" event={"ID":"d5ee0a49-193d-478a-af87-42c6fb7ea091","Type":"ContainerStarted","Data":"7e186b4c85e9255c76cf809801372380f84684432210be646644751017745c85"} Feb 20 09:27:02 crc kubenswrapper[4492]: I0220 09:27:02.172711 4492 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 20 09:27:03 crc kubenswrapper[4492]: I0220 09:27:03.565233 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:27:03 crc kubenswrapper[4492]: E0220 09:27:03.565713 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:27:04 crc kubenswrapper[4492]: I0220 09:27:04.189809 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bltlg" event={"ID":"d5ee0a49-193d-478a-af87-42c6fb7ea091","Type":"ContainerStarted","Data":"89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e"} Feb 20 09:27:07 crc kubenswrapper[4492]: I0220 09:27:07.219823 4492 generic.go:334] "Generic (PLEG): container finished" podID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerID="89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e" exitCode=0 Feb 20 09:27:07 crc kubenswrapper[4492]: I0220 09:27:07.220352 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bltlg" event={"ID":"d5ee0a49-193d-478a-af87-42c6fb7ea091","Type":"ContainerDied","Data":"89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e"} Feb 20 09:27:08 crc kubenswrapper[4492]: I0220 09:27:08.231986 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bltlg" event={"ID":"d5ee0a49-193d-478a-af87-42c6fb7ea091","Type":"ContainerStarted","Data":"6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872"} Feb 20 09:27:10 crc kubenswrapper[4492]: I0220 09:27:10.830939 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:10 crc kubenswrapper[4492]: I0220 09:27:10.832180 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:11 crc kubenswrapper[4492]: I0220 09:27:11.867777 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bltlg" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="registry-server" probeResult="failure" output=< Feb 20 09:27:11 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:27:11 crc kubenswrapper[4492]: > Feb 20 09:27:18 crc kubenswrapper[4492]: I0220 09:27:18.557114 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:27:18 crc kubenswrapper[4492]: E0220 09:27:18.558277 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:27:21 crc kubenswrapper[4492]: I0220 09:27:21.875773 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bltlg" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="registry-server" probeResult="failure" output=< Feb 20 09:27:21 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:27:21 crc kubenswrapper[4492]: > Feb 20 09:27:30 crc kubenswrapper[4492]: I0220 09:27:30.557444 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:27:30 crc kubenswrapper[4492]: E0220 09:27:30.558443 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:27:30 crc kubenswrapper[4492]: I0220 09:27:30.875291 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:30 crc kubenswrapper[4492]: I0220 09:27:30.906850 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bltlg" podStartSLOduration=25.291025546 podStartE2EDuration="30.903461749s" podCreationTimestamp="2026-02-20 09:27:00 +0000 UTC" firstStartedPulling="2026-02-20 09:27:02.170773476 +0000 UTC m=+9978.942062444" lastFinishedPulling="2026-02-20 09:27:07.78320967 +0000 UTC m=+9984.554498647" observedRunningTime="2026-02-20 09:27:08.258209223 +0000 UTC m=+9985.029498201" watchObservedRunningTime="2026-02-20 09:27:30.903461749 +0000 UTC m=+10007.674750757" Feb 20 09:27:30 crc kubenswrapper[4492]: I0220 09:27:30.932454 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:31 crc kubenswrapper[4492]: I0220 09:27:31.693523 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bltlg"] Feb 20 09:27:32 crc kubenswrapper[4492]: I0220 09:27:32.471367 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bltlg" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="registry-server" containerID="cri-o://6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872" gracePeriod=2 Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.069439 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.081644 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-catalog-content\") pod \"d5ee0a49-193d-478a-af87-42c6fb7ea091\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.081705 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-utilities\") pod \"d5ee0a49-193d-478a-af87-42c6fb7ea091\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.082023 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7x7c\" (UniqueName: \"kubernetes.io/projected/d5ee0a49-193d-478a-af87-42c6fb7ea091-kube-api-access-g7x7c\") pod \"d5ee0a49-193d-478a-af87-42c6fb7ea091\" (UID: \"d5ee0a49-193d-478a-af87-42c6fb7ea091\") " Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.083351 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-utilities" (OuterVolumeSpecName: "utilities") pod "d5ee0a49-193d-478a-af87-42c6fb7ea091" (UID: "d5ee0a49-193d-478a-af87-42c6fb7ea091"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.098209 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5ee0a49-193d-478a-af87-42c6fb7ea091-kube-api-access-g7x7c" (OuterVolumeSpecName: "kube-api-access-g7x7c") pod "d5ee0a49-193d-478a-af87-42c6fb7ea091" (UID: "d5ee0a49-193d-478a-af87-42c6fb7ea091"). InnerVolumeSpecName "kube-api-access-g7x7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.167492 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5ee0a49-193d-478a-af87-42c6fb7ea091" (UID: "d5ee0a49-193d-478a-af87-42c6fb7ea091"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.184302 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7x7c\" (UniqueName: \"kubernetes.io/projected/d5ee0a49-193d-478a-af87-42c6fb7ea091-kube-api-access-g7x7c\") on node \"crc\" DevicePath \"\"" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.184330 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.184354 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5ee0a49-193d-478a-af87-42c6fb7ea091-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.480186 4492 generic.go:334] "Generic (PLEG): container finished" podID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerID="6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872" exitCode=0 Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.480256 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bltlg" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.480299 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bltlg" event={"ID":"d5ee0a49-193d-478a-af87-42c6fb7ea091","Type":"ContainerDied","Data":"6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872"} Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.480885 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bltlg" event={"ID":"d5ee0a49-193d-478a-af87-42c6fb7ea091","Type":"ContainerDied","Data":"7e186b4c85e9255c76cf809801372380f84684432210be646644751017745c85"} Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.480909 4492 scope.go:117] "RemoveContainer" containerID="6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.520615 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bltlg"] Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.524100 4492 scope.go:117] "RemoveContainer" containerID="89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.526648 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bltlg"] Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.555789 4492 scope.go:117] "RemoveContainer" containerID="f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.583912 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" path="/var/lib/kubelet/pods/d5ee0a49-193d-478a-af87-42c6fb7ea091/volumes" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.600253 4492 scope.go:117] "RemoveContainer" containerID="6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872" Feb 20 09:27:33 crc kubenswrapper[4492]: E0220 09:27:33.603497 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872\": container with ID starting with 6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872 not found: ID does not exist" containerID="6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.603571 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872"} err="failed to get container status \"6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872\": rpc error: code = NotFound desc = could not find container \"6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872\": container with ID starting with 6ec219d7e1c6122a5a71520c8994fed57c777f0670ddb7a716bb882476a08872 not found: ID does not exist" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.603604 4492 scope.go:117] "RemoveContainer" containerID="89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e" Feb 20 09:27:33 crc kubenswrapper[4492]: E0220 09:27:33.605688 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e\": container with ID starting with 89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e not found: ID does not exist" containerID="89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.605894 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e"} err="failed to get container status \"89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e\": rpc error: code = NotFound desc = could not find container \"89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e\": container with ID starting with 89db0ef72f8185c0fdd5d580d77424145a6314cf7890f2b47ed04020db3aaa0e not found: ID does not exist" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.606011 4492 scope.go:117] "RemoveContainer" containerID="f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80" Feb 20 09:27:33 crc kubenswrapper[4492]: E0220 09:27:33.606507 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80\": container with ID starting with f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80 not found: ID does not exist" containerID="f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80" Feb 20 09:27:33 crc kubenswrapper[4492]: I0220 09:27:33.606540 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80"} err="failed to get container status \"f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80\": rpc error: code = NotFound desc = could not find container \"f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80\": container with ID starting with f374387c5aaf3222e24042a8c0c6e43615729309e9990c405ad9742376821d80 not found: ID does not exist" Feb 20 09:27:43 crc kubenswrapper[4492]: I0220 09:27:43.563459 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:27:43 crc kubenswrapper[4492]: E0220 09:27:43.565397 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:27:55 crc kubenswrapper[4492]: I0220 09:27:55.556855 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:27:55 crc kubenswrapper[4492]: E0220 09:27:55.557879 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:28:07 crc kubenswrapper[4492]: I0220 09:28:07.557079 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:28:07 crc kubenswrapper[4492]: E0220 09:28:07.558030 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:28:11 crc kubenswrapper[4492]: I0220 09:28:11.640055 4492 scope.go:117] "RemoveContainer" containerID="f78c0eb0a607de08007736cd2f9ad301dcd228413bf9a1a39588a24f365eb08f" Feb 20 09:28:21 crc kubenswrapper[4492]: I0220 09:28:21.558742 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:28:21 crc kubenswrapper[4492]: E0220 09:28:21.559618 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:28:28 crc kubenswrapper[4492]: I0220 09:28:28.983070 4492 generic.go:334] "Generic (PLEG): container finished" podID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerID="7029bca25b42ac67a0250597d1aaab172eee7f40f68b5548fdc3ad9f759c23cf" exitCode=0 Feb 20 09:28:28 crc kubenswrapper[4492]: I0220 09:28:28.983164 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7grgm/must-gather-rdc6g" event={"ID":"37384967-b137-4cdf-868d-74cb7dfcc85c","Type":"ContainerDied","Data":"7029bca25b42ac67a0250597d1aaab172eee7f40f68b5548fdc3ad9f759c23cf"} Feb 20 09:28:28 crc kubenswrapper[4492]: I0220 09:28:28.985040 4492 scope.go:117] "RemoveContainer" containerID="7029bca25b42ac67a0250597d1aaab172eee7f40f68b5548fdc3ad9f759c23cf" Feb 20 09:28:29 crc kubenswrapper[4492]: I0220 09:28:29.376705 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7grgm_must-gather-rdc6g_37384967-b137-4cdf-868d-74cb7dfcc85c/gather/0.log" Feb 20 09:28:34 crc kubenswrapper[4492]: E0220 09:28:34.204236 4492 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.25:40128->192.168.26.25:44007: write tcp 192.168.26.25:40128->192.168.26.25:44007: write: broken pipe Feb 20 09:28:34 crc kubenswrapper[4492]: I0220 09:28:34.556800 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:28:34 crc kubenswrapper[4492]: E0220 09:28:34.557174 4492 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zrlsb_openshift-machine-config-operator(f8755354-8581-4b5d-9b7e-6b7c53707f8c)\"" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" Feb 20 09:28:39 crc kubenswrapper[4492]: I0220 09:28:39.718820 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7grgm/must-gather-rdc6g"] Feb 20 09:28:39 crc kubenswrapper[4492]: I0220 09:28:39.725652 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7grgm/must-gather-rdc6g" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerName="copy" containerID="cri-o://358fec6ee7a399344302d202cfd505b965459adc45196c72c3ba135472c18b65" gracePeriod=2 Feb 20 09:28:39 crc kubenswrapper[4492]: I0220 09:28:39.754799 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7grgm/must-gather-rdc6g"] Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.123950 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7grgm_must-gather-rdc6g_37384967-b137-4cdf-868d-74cb7dfcc85c/copy/0.log" Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.124764 4492 generic.go:334] "Generic (PLEG): container finished" podID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerID="358fec6ee7a399344302d202cfd505b965459adc45196c72c3ba135472c18b65" exitCode=143 Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.367272 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7grgm_must-gather-rdc6g_37384967-b137-4cdf-868d-74cb7dfcc85c/copy/0.log" Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.368584 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.481351 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/37384967-b137-4cdf-868d-74cb7dfcc85c-must-gather-output\") pod \"37384967-b137-4cdf-868d-74cb7dfcc85c\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.481772 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sqt7\" (UniqueName: \"kubernetes.io/projected/37384967-b137-4cdf-868d-74cb7dfcc85c-kube-api-access-8sqt7\") pod \"37384967-b137-4cdf-868d-74cb7dfcc85c\" (UID: \"37384967-b137-4cdf-868d-74cb7dfcc85c\") " Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.506235 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37384967-b137-4cdf-868d-74cb7dfcc85c-kube-api-access-8sqt7" (OuterVolumeSpecName: "kube-api-access-8sqt7") pod "37384967-b137-4cdf-868d-74cb7dfcc85c" (UID: "37384967-b137-4cdf-868d-74cb7dfcc85c"). InnerVolumeSpecName "kube-api-access-8sqt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.585238 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sqt7\" (UniqueName: \"kubernetes.io/projected/37384967-b137-4cdf-868d-74cb7dfcc85c-kube-api-access-8sqt7\") on node \"crc\" DevicePath \"\"" Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.619099 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37384967-b137-4cdf-868d-74cb7dfcc85c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "37384967-b137-4cdf-868d-74cb7dfcc85c" (UID: "37384967-b137-4cdf-868d-74cb7dfcc85c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:28:40 crc kubenswrapper[4492]: I0220 09:28:40.689047 4492 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/37384967-b137-4cdf-868d-74cb7dfcc85c-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 20 09:28:41 crc kubenswrapper[4492]: I0220 09:28:41.134498 4492 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7grgm_must-gather-rdc6g_37384967-b137-4cdf-868d-74cb7dfcc85c/copy/0.log" Feb 20 09:28:41 crc kubenswrapper[4492]: I0220 09:28:41.137014 4492 scope.go:117] "RemoveContainer" containerID="358fec6ee7a399344302d202cfd505b965459adc45196c72c3ba135472c18b65" Feb 20 09:28:41 crc kubenswrapper[4492]: I0220 09:28:41.137069 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7grgm/must-gather-rdc6g" Feb 20 09:28:41 crc kubenswrapper[4492]: I0220 09:28:41.178840 4492 scope.go:117] "RemoveContainer" containerID="7029bca25b42ac67a0250597d1aaab172eee7f40f68b5548fdc3ad9f759c23cf" Feb 20 09:28:41 crc kubenswrapper[4492]: I0220 09:28:41.568565 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" path="/var/lib/kubelet/pods/37384967-b137-4cdf-868d-74cb7dfcc85c/volumes" Feb 20 09:28:46 crc kubenswrapper[4492]: I0220 09:28:46.557800 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" Feb 20 09:28:47 crc kubenswrapper[4492]: I0220 09:28:47.196986 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"fbaac528affb82db28220a7361d65e6c7d8198f9422e8202431bdfff10da5453"} Feb 20 09:29:11 crc kubenswrapper[4492]: I0220 09:29:11.764399 4492 scope.go:117] "RemoveContainer" containerID="2d7dd333ce0014f99fb6ef295945aacd68033f536f4f8ae56477187165135221" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.210538 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm"] Feb 20 09:30:00 crc kubenswrapper[4492]: E0220 09:30:00.214130 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="registry-server" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.214156 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="registry-server" Feb 20 09:30:00 crc kubenswrapper[4492]: E0220 09:30:00.214185 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerName="gather" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.214193 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerName="gather" Feb 20 09:30:00 crc kubenswrapper[4492]: E0220 09:30:00.214208 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="extract-content" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.214214 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="extract-content" Feb 20 09:30:00 crc kubenswrapper[4492]: E0220 09:30:00.214228 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="extract-utilities" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.214234 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="extract-utilities" Feb 20 09:30:00 crc kubenswrapper[4492]: E0220 09:30:00.214254 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerName="copy" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.214260 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerName="copy" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.215497 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerName="copy" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.215648 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5ee0a49-193d-478a-af87-42c6fb7ea091" containerName="registry-server" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.215673 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="37384967-b137-4cdf-868d-74cb7dfcc85c" containerName="gather" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.218558 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.226247 4492 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.226249 4492 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.229060 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm"] Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.236075 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg5rs\" (UniqueName: \"kubernetes.io/projected/64ac0f07-8817-45e4-9118-12af28d41677-kube-api-access-lg5rs\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.236163 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64ac0f07-8817-45e4-9118-12af28d41677-config-volume\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.236199 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64ac0f07-8817-45e4-9118-12af28d41677-secret-volume\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.338779 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64ac0f07-8817-45e4-9118-12af28d41677-config-volume\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.338921 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64ac0f07-8817-45e4-9118-12af28d41677-secret-volume\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.339199 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg5rs\" (UniqueName: \"kubernetes.io/projected/64ac0f07-8817-45e4-9118-12af28d41677-kube-api-access-lg5rs\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.339788 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64ac0f07-8817-45e4-9118-12af28d41677-config-volume\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.347061 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64ac0f07-8817-45e4-9118-12af28d41677-secret-volume\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.354601 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg5rs\" (UniqueName: \"kubernetes.io/projected/64ac0f07-8817-45e4-9118-12af28d41677-kube-api-access-lg5rs\") pod \"collect-profiles-29526330-rfwrm\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:00 crc kubenswrapper[4492]: I0220 09:30:00.541711 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:01 crc kubenswrapper[4492]: I0220 09:30:01.233802 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm"] Feb 20 09:30:01 crc kubenswrapper[4492]: W0220 09:30:01.240970 4492 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64ac0f07_8817_45e4_9118_12af28d41677.slice/crio-00050173f6572c0e3a338cea60baac9b6507fd1fd0a05683f4865cdfbaf2bc60 WatchSource:0}: Error finding container 00050173f6572c0e3a338cea60baac9b6507fd1fd0a05683f4865cdfbaf2bc60: Status 404 returned error can't find the container with id 00050173f6572c0e3a338cea60baac9b6507fd1fd0a05683f4865cdfbaf2bc60 Feb 20 09:30:01 crc kubenswrapper[4492]: I0220 09:30:01.979006 4492 generic.go:334] "Generic (PLEG): container finished" podID="64ac0f07-8817-45e4-9118-12af28d41677" containerID="a259e24c2f150d6e9304f3a0d4199119a7dcd77c67c45cd6edbc14032a332f1d" exitCode=0 Feb 20 09:30:01 crc kubenswrapper[4492]: I0220 09:30:01.979368 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" event={"ID":"64ac0f07-8817-45e4-9118-12af28d41677","Type":"ContainerDied","Data":"a259e24c2f150d6e9304f3a0d4199119a7dcd77c67c45cd6edbc14032a332f1d"} Feb 20 09:30:01 crc kubenswrapper[4492]: I0220 09:30:01.979404 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" event={"ID":"64ac0f07-8817-45e4-9118-12af28d41677","Type":"ContainerStarted","Data":"00050173f6572c0e3a338cea60baac9b6507fd1fd0a05683f4865cdfbaf2bc60"} Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.254716 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.304676 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg5rs\" (UniqueName: \"kubernetes.io/projected/64ac0f07-8817-45e4-9118-12af28d41677-kube-api-access-lg5rs\") pod \"64ac0f07-8817-45e4-9118-12af28d41677\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.313734 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64ac0f07-8817-45e4-9118-12af28d41677-kube-api-access-lg5rs" (OuterVolumeSpecName: "kube-api-access-lg5rs") pod "64ac0f07-8817-45e4-9118-12af28d41677" (UID: "64ac0f07-8817-45e4-9118-12af28d41677"). InnerVolumeSpecName "kube-api-access-lg5rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.405770 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64ac0f07-8817-45e4-9118-12af28d41677-secret-volume\") pod \"64ac0f07-8817-45e4-9118-12af28d41677\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.405814 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64ac0f07-8817-45e4-9118-12af28d41677-config-volume\") pod \"64ac0f07-8817-45e4-9118-12af28d41677\" (UID: \"64ac0f07-8817-45e4-9118-12af28d41677\") " Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.406143 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg5rs\" (UniqueName: \"kubernetes.io/projected/64ac0f07-8817-45e4-9118-12af28d41677-kube-api-access-lg5rs\") on node \"crc\" DevicePath \"\"" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.406412 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64ac0f07-8817-45e4-9118-12af28d41677-config-volume" (OuterVolumeSpecName: "config-volume") pod "64ac0f07-8817-45e4-9118-12af28d41677" (UID: "64ac0f07-8817-45e4-9118-12af28d41677"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.409934 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64ac0f07-8817-45e4-9118-12af28d41677-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "64ac0f07-8817-45e4-9118-12af28d41677" (UID: "64ac0f07-8817-45e4-9118-12af28d41677"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.508506 4492 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/64ac0f07-8817-45e4-9118-12af28d41677-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.508531 4492 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/64ac0f07-8817-45e4-9118-12af28d41677-config-volume\") on node \"crc\" DevicePath \"\"" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.998659 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" event={"ID":"64ac0f07-8817-45e4-9118-12af28d41677","Type":"ContainerDied","Data":"00050173f6572c0e3a338cea60baac9b6507fd1fd0a05683f4865cdfbaf2bc60"} Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.999079 4492 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00050173f6572c0e3a338cea60baac9b6507fd1fd0a05683f4865cdfbaf2bc60" Feb 20 09:30:03 crc kubenswrapper[4492]: I0220 09:30:03.998742 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29526330-rfwrm" Feb 20 09:30:04 crc kubenswrapper[4492]: I0220 09:30:04.332286 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr"] Feb 20 09:30:04 crc kubenswrapper[4492]: I0220 09:30:04.343850 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29526285-f58xr"] Feb 20 09:30:05 crc kubenswrapper[4492]: I0220 09:30:05.567508 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d036832f-1f0c-4f37-be93-5173bc06e81f" path="/var/lib/kubelet/pods/d036832f-1f0c-4f37-be93-5173bc06e81f/volumes" Feb 20 09:30:11 crc kubenswrapper[4492]: I0220 09:30:11.856724 4492 scope.go:117] "RemoveContainer" containerID="37e4936de6f5566e8270e0e4b05fca9f2e01d7aaff12d46661ba239ac660d20c" Feb 20 09:30:14 crc kubenswrapper[4492]: I0220 09:30:14.934683 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hw6c9"] Feb 20 09:30:14 crc kubenswrapper[4492]: E0220 09:30:14.935390 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64ac0f07-8817-45e4-9118-12af28d41677" containerName="collect-profiles" Feb 20 09:30:14 crc kubenswrapper[4492]: I0220 09:30:14.935404 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="64ac0f07-8817-45e4-9118-12af28d41677" containerName="collect-profiles" Feb 20 09:30:14 crc kubenswrapper[4492]: I0220 09:30:14.935676 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="64ac0f07-8817-45e4-9118-12af28d41677" containerName="collect-profiles" Feb 20 09:30:14 crc kubenswrapper[4492]: I0220 09:30:14.937197 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:14 crc kubenswrapper[4492]: I0220 09:30:14.943854 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hw6c9"] Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.034408 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fa234-15ae-4257-8d57-b10d65cf34c8-catalog-content\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.034689 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fa234-15ae-4257-8d57-b10d65cf34c8-utilities\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.034751 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4grjn\" (UniqueName: \"kubernetes.io/projected/2d2fa234-15ae-4257-8d57-b10d65cf34c8-kube-api-access-4grjn\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.136577 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4grjn\" (UniqueName: \"kubernetes.io/projected/2d2fa234-15ae-4257-8d57-b10d65cf34c8-kube-api-access-4grjn\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.136688 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fa234-15ae-4257-8d57-b10d65cf34c8-catalog-content\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.136957 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fa234-15ae-4257-8d57-b10d65cf34c8-utilities\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.137435 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fa234-15ae-4257-8d57-b10d65cf34c8-utilities\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.137617 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fa234-15ae-4257-8d57-b10d65cf34c8-catalog-content\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.162158 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4grjn\" (UniqueName: \"kubernetes.io/projected/2d2fa234-15ae-4257-8d57-b10d65cf34c8-kube-api-access-4grjn\") pod \"certified-operators-hw6c9\" (UID: \"2d2fa234-15ae-4257-8d57-b10d65cf34c8\") " pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.255110 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:15 crc kubenswrapper[4492]: I0220 09:30:15.671596 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hw6c9"] Feb 20 09:30:16 crc kubenswrapper[4492]: I0220 09:30:16.095084 4492 generic.go:334] "Generic (PLEG): container finished" podID="2d2fa234-15ae-4257-8d57-b10d65cf34c8" containerID="fc541d6e1f5e50b902072ddc0716accb2ac04c4977214ac3f054621fc5251944" exitCode=0 Feb 20 09:30:16 crc kubenswrapper[4492]: I0220 09:30:16.095117 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hw6c9" event={"ID":"2d2fa234-15ae-4257-8d57-b10d65cf34c8","Type":"ContainerDied","Data":"fc541d6e1f5e50b902072ddc0716accb2ac04c4977214ac3f054621fc5251944"} Feb 20 09:30:16 crc kubenswrapper[4492]: I0220 09:30:16.095421 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hw6c9" event={"ID":"2d2fa234-15ae-4257-8d57-b10d65cf34c8","Type":"ContainerStarted","Data":"3f4f122102c21af9cae5391a5f27e81206814655cd269bb18dfbc5e37c88a96e"} Feb 20 09:30:21 crc kubenswrapper[4492]: I0220 09:30:21.140179 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hw6c9" event={"ID":"2d2fa234-15ae-4257-8d57-b10d65cf34c8","Type":"ContainerStarted","Data":"364ee958e1b18271bcbc77c9b8c07e54d8c89b559bd46e329f774870bb0e8d06"} Feb 20 09:30:22 crc kubenswrapper[4492]: I0220 09:30:22.151994 4492 generic.go:334] "Generic (PLEG): container finished" podID="2d2fa234-15ae-4257-8d57-b10d65cf34c8" containerID="364ee958e1b18271bcbc77c9b8c07e54d8c89b559bd46e329f774870bb0e8d06" exitCode=0 Feb 20 09:30:22 crc kubenswrapper[4492]: I0220 09:30:22.152073 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hw6c9" event={"ID":"2d2fa234-15ae-4257-8d57-b10d65cf34c8","Type":"ContainerDied","Data":"364ee958e1b18271bcbc77c9b8c07e54d8c89b559bd46e329f774870bb0e8d06"} Feb 20 09:30:23 crc kubenswrapper[4492]: I0220 09:30:23.163962 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hw6c9" event={"ID":"2d2fa234-15ae-4257-8d57-b10d65cf34c8","Type":"ContainerStarted","Data":"ea0fd9a2b49887d8f66b24bb84485e1bd53dbd9b06d777deb148d84983fa7cea"} Feb 20 09:30:23 crc kubenswrapper[4492]: I0220 09:30:23.190251 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hw6c9" podStartSLOduration=2.6561288899999997 podStartE2EDuration="9.18889687s" podCreationTimestamp="2026-02-20 09:30:14 +0000 UTC" firstStartedPulling="2026-02-20 09:30:16.097311372 +0000 UTC m=+10172.868600350" lastFinishedPulling="2026-02-20 09:30:22.630079352 +0000 UTC m=+10179.401368330" observedRunningTime="2026-02-20 09:30:23.179962263 +0000 UTC m=+10179.951251242" watchObservedRunningTime="2026-02-20 09:30:23.18889687 +0000 UTC m=+10179.960185849" Feb 20 09:30:25 crc kubenswrapper[4492]: I0220 09:30:25.256791 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:25 crc kubenswrapper[4492]: I0220 09:30:25.257049 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:26 crc kubenswrapper[4492]: I0220 09:30:26.288458 4492 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-hw6c9" podUID="2d2fa234-15ae-4257-8d57-b10d65cf34c8" containerName="registry-server" probeResult="failure" output=< Feb 20 09:30:26 crc kubenswrapper[4492]: timeout: failed to connect service ":50051" within 1s Feb 20 09:30:26 crc kubenswrapper[4492]: > Feb 20 09:30:35 crc kubenswrapper[4492]: I0220 09:30:35.296099 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:35 crc kubenswrapper[4492]: I0220 09:30:35.335021 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hw6c9" Feb 20 09:30:35 crc kubenswrapper[4492]: I0220 09:30:35.389269 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hw6c9"] Feb 20 09:30:35 crc kubenswrapper[4492]: I0220 09:30:35.529735 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vc5ws"] Feb 20 09:30:35 crc kubenswrapper[4492]: I0220 09:30:35.532383 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vc5ws" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="registry-server" containerID="cri-o://eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee" gracePeriod=2 Feb 20 09:30:35 crc kubenswrapper[4492]: I0220 09:30:35.966029 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.027543 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-catalog-content\") pod \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.027651 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgxzk\" (UniqueName: \"kubernetes.io/projected/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-kube-api-access-qgxzk\") pod \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.027773 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-utilities\") pod \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\" (UID: \"b0e828a1-a902-411f-8b55-f12b3a8c3e1c\") " Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.028838 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-utilities" (OuterVolumeSpecName: "utilities") pod "b0e828a1-a902-411f-8b55-f12b3a8c3e1c" (UID: "b0e828a1-a902-411f-8b55-f12b3a8c3e1c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.082052 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0e828a1-a902-411f-8b55-f12b3a8c3e1c" (UID: "b0e828a1-a902-411f-8b55-f12b3a8c3e1c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.129856 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.129890 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.287210 4492 generic.go:334] "Generic (PLEG): container finished" podID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerID="eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee" exitCode=0 Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.287390 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vc5ws" event={"ID":"b0e828a1-a902-411f-8b55-f12b3a8c3e1c","Type":"ContainerDied","Data":"eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee"} Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.287442 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vc5ws" event={"ID":"b0e828a1-a902-411f-8b55-f12b3a8c3e1c","Type":"ContainerDied","Data":"ae375b830ddb6e16c922c8f93a22e0a1528e077c01fafc00fb609da80c36acbf"} Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.287463 4492 scope.go:117] "RemoveContainer" containerID="eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.287556 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vc5ws" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.311571 4492 scope.go:117] "RemoveContainer" containerID="c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.407954 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-kube-api-access-qgxzk" (OuterVolumeSpecName: "kube-api-access-qgxzk") pod "b0e828a1-a902-411f-8b55-f12b3a8c3e1c" (UID: "b0e828a1-a902-411f-8b55-f12b3a8c3e1c"). InnerVolumeSpecName "kube-api-access-qgxzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.419052 4492 scope.go:117] "RemoveContainer" containerID="91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.435456 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgxzk\" (UniqueName: \"kubernetes.io/projected/b0e828a1-a902-411f-8b55-f12b3a8c3e1c-kube-api-access-qgxzk\") on node \"crc\" DevicePath \"\"" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.495658 4492 scope.go:117] "RemoveContainer" containerID="eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee" Feb 20 09:30:36 crc kubenswrapper[4492]: E0220 09:30:36.497414 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee\": container with ID starting with eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee not found: ID does not exist" containerID="eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.497466 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee"} err="failed to get container status \"eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee\": rpc error: code = NotFound desc = could not find container \"eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee\": container with ID starting with eaa7db644a8d41c066b3955caa51e12c2541158b97d8f24c69b0ca378387fdee not found: ID does not exist" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.497508 4492 scope.go:117] "RemoveContainer" containerID="c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e" Feb 20 09:30:36 crc kubenswrapper[4492]: E0220 09:30:36.497843 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e\": container with ID starting with c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e not found: ID does not exist" containerID="c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.497863 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e"} err="failed to get container status \"c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e\": rpc error: code = NotFound desc = could not find container \"c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e\": container with ID starting with c3f85ff7e8e4a88077fb45e7766f0019cd39aca2876a99521177808c14ae6f6e not found: ID does not exist" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.497878 4492 scope.go:117] "RemoveContainer" containerID="91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68" Feb 20 09:30:36 crc kubenswrapper[4492]: E0220 09:30:36.498132 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68\": container with ID starting with 91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68 not found: ID does not exist" containerID="91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.498155 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68"} err="failed to get container status \"91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68\": rpc error: code = NotFound desc = could not find container \"91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68\": container with ID starting with 91a54a9b8e55f05a76ab2b73895cd22d9cace494cc410520c564ac25fcf51c68 not found: ID does not exist" Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.616397 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vc5ws"] Feb 20 09:30:36 crc kubenswrapper[4492]: I0220 09:30:36.623819 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vc5ws"] Feb 20 09:30:37 crc kubenswrapper[4492]: I0220 09:30:37.577119 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" path="/var/lib/kubelet/pods/b0e828a1-a902-411f-8b55-f12b3a8c3e1c/volumes" Feb 20 09:31:09 crc kubenswrapper[4492]: I0220 09:31:09.311345 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:31:09 crc kubenswrapper[4492]: I0220 09:31:09.311986 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:31:39 crc kubenswrapper[4492]: I0220 09:31:39.311169 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:31:39 crc kubenswrapper[4492]: I0220 09:31:39.311975 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.811563 4492 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jbzmt"] Feb 20 09:31:54 crc kubenswrapper[4492]: E0220 09:31:54.814130 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="registry-server" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.814170 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="registry-server" Feb 20 09:31:54 crc kubenswrapper[4492]: E0220 09:31:54.814228 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="extract-content" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.814235 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="extract-content" Feb 20 09:31:54 crc kubenswrapper[4492]: E0220 09:31:54.814254 4492 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="extract-utilities" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.814263 4492 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="extract-utilities" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.814868 4492 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0e828a1-a902-411f-8b55-f12b3a8c3e1c" containerName="registry-server" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.818984 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.868954 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbzmt"] Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.910494 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss7x7\" (UniqueName: \"kubernetes.io/projected/1121159f-d018-464b-a012-afad30e73f00-kube-api-access-ss7x7\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.910564 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-utilities\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:54 crc kubenswrapper[4492]: I0220 09:31:54.910635 4492 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-catalog-content\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.014045 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss7x7\" (UniqueName: \"kubernetes.io/projected/1121159f-d018-464b-a012-afad30e73f00-kube-api-access-ss7x7\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.014103 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-utilities\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.014168 4492 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-catalog-content\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.014672 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-utilities\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.014746 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-catalog-content\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.041695 4492 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss7x7\" (UniqueName: \"kubernetes.io/projected/1121159f-d018-464b-a012-afad30e73f00-kube-api-access-ss7x7\") pod \"community-operators-jbzmt\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.149840 4492 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.691352 4492 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbzmt"] Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.949089 4492 generic.go:334] "Generic (PLEG): container finished" podID="1121159f-d018-464b-a012-afad30e73f00" containerID="54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b" exitCode=0 Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.949257 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzmt" event={"ID":"1121159f-d018-464b-a012-afad30e73f00","Type":"ContainerDied","Data":"54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b"} Feb 20 09:31:55 crc kubenswrapper[4492]: I0220 09:31:55.949461 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzmt" event={"ID":"1121159f-d018-464b-a012-afad30e73f00","Type":"ContainerStarted","Data":"17fbe1bc7dc7fbb32c9b250ae1b4b4cde3373bfdf38eb45b5001bab07db47ef2"} Feb 20 09:31:57 crc kubenswrapper[4492]: I0220 09:31:57.971442 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzmt" event={"ID":"1121159f-d018-464b-a012-afad30e73f00","Type":"ContainerStarted","Data":"16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb"} Feb 20 09:31:58 crc kubenswrapper[4492]: I0220 09:31:58.984989 4492 generic.go:334] "Generic (PLEG): container finished" podID="1121159f-d018-464b-a012-afad30e73f00" containerID="16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb" exitCode=0 Feb 20 09:31:58 crc kubenswrapper[4492]: I0220 09:31:58.985178 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzmt" event={"ID":"1121159f-d018-464b-a012-afad30e73f00","Type":"ContainerDied","Data":"16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb"} Feb 20 09:32:00 crc kubenswrapper[4492]: I0220 09:32:00.002517 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzmt" event={"ID":"1121159f-d018-464b-a012-afad30e73f00","Type":"ContainerStarted","Data":"f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd"} Feb 20 09:32:00 crc kubenswrapper[4492]: I0220 09:32:00.029098 4492 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jbzmt" podStartSLOduration=2.547201672 podStartE2EDuration="6.029081068s" podCreationTimestamp="2026-02-20 09:31:54 +0000 UTC" firstStartedPulling="2026-02-20 09:31:55.952051726 +0000 UTC m=+10272.723340705" lastFinishedPulling="2026-02-20 09:31:59.433931123 +0000 UTC m=+10276.205220101" observedRunningTime="2026-02-20 09:32:00.024535417 +0000 UTC m=+10276.795824395" watchObservedRunningTime="2026-02-20 09:32:00.029081068 +0000 UTC m=+10276.800370035" Feb 20 09:32:05 crc kubenswrapper[4492]: I0220 09:32:05.150463 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:32:05 crc kubenswrapper[4492]: I0220 09:32:05.151178 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:32:05 crc kubenswrapper[4492]: I0220 09:32:05.193708 4492 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:32:06 crc kubenswrapper[4492]: I0220 09:32:06.236751 4492 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:32:06 crc kubenswrapper[4492]: I0220 09:32:06.295346 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbzmt"] Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.076498 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jbzmt" podUID="1121159f-d018-464b-a012-afad30e73f00" containerName="registry-server" containerID="cri-o://f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd" gracePeriod=2 Feb 20 09:32:08 crc kubenswrapper[4492]: E0220 09:32:08.321671 4492 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1121159f_d018_464b_a012_afad30e73f00.slice/crio-conmon-f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1121159f_d018_464b_a012_afad30e73f00.slice/crio-f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd.scope\": RecentStats: unable to find data in memory cache]" Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.676215 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.809165 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-utilities\") pod \"1121159f-d018-464b-a012-afad30e73f00\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.809237 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss7x7\" (UniqueName: \"kubernetes.io/projected/1121159f-d018-464b-a012-afad30e73f00-kube-api-access-ss7x7\") pod \"1121159f-d018-464b-a012-afad30e73f00\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.809406 4492 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-catalog-content\") pod \"1121159f-d018-464b-a012-afad30e73f00\" (UID: \"1121159f-d018-464b-a012-afad30e73f00\") " Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.810122 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-utilities" (OuterVolumeSpecName: "utilities") pod "1121159f-d018-464b-a012-afad30e73f00" (UID: "1121159f-d018-464b-a012-afad30e73f00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.820793 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1121159f-d018-464b-a012-afad30e73f00-kube-api-access-ss7x7" (OuterVolumeSpecName: "kube-api-access-ss7x7") pod "1121159f-d018-464b-a012-afad30e73f00" (UID: "1121159f-d018-464b-a012-afad30e73f00"). InnerVolumeSpecName "kube-api-access-ss7x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.849652 4492 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1121159f-d018-464b-a012-afad30e73f00" (UID: "1121159f-d018-464b-a012-afad30e73f00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.912348 4492 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.912387 4492 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1121159f-d018-464b-a012-afad30e73f00-utilities\") on node \"crc\" DevicePath \"\"" Feb 20 09:32:08 crc kubenswrapper[4492]: I0220 09:32:08.912400 4492 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss7x7\" (UniqueName: \"kubernetes.io/projected/1121159f-d018-464b-a012-afad30e73f00-kube-api-access-ss7x7\") on node \"crc\" DevicePath \"\"" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.088198 4492 generic.go:334] "Generic (PLEG): container finished" podID="1121159f-d018-464b-a012-afad30e73f00" containerID="f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd" exitCode=0 Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.088244 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzmt" event={"ID":"1121159f-d018-464b-a012-afad30e73f00","Type":"ContainerDied","Data":"f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd"} Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.088250 4492 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbzmt" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.088273 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbzmt" event={"ID":"1121159f-d018-464b-a012-afad30e73f00","Type":"ContainerDied","Data":"17fbe1bc7dc7fbb32c9b250ae1b4b4cde3373bfdf38eb45b5001bab07db47ef2"} Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.088295 4492 scope.go:117] "RemoveContainer" containerID="f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.117319 4492 scope.go:117] "RemoveContainer" containerID="16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.122593 4492 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbzmt"] Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.133250 4492 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jbzmt"] Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.137623 4492 scope.go:117] "RemoveContainer" containerID="54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.173549 4492 scope.go:117] "RemoveContainer" containerID="f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd" Feb 20 09:32:09 crc kubenswrapper[4492]: E0220 09:32:09.173988 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd\": container with ID starting with f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd not found: ID does not exist" containerID="f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.174075 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd"} err="failed to get container status \"f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd\": rpc error: code = NotFound desc = could not find container \"f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd\": container with ID starting with f92215d72553438881bb515eeb9965e4a77c0a4c920087a83eb4e75bf08821dd not found: ID does not exist" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.174158 4492 scope.go:117] "RemoveContainer" containerID="16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb" Feb 20 09:32:09 crc kubenswrapper[4492]: E0220 09:32:09.174456 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb\": container with ID starting with 16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb not found: ID does not exist" containerID="16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.174549 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb"} err="failed to get container status \"16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb\": rpc error: code = NotFound desc = could not find container \"16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb\": container with ID starting with 16d52f62df61819da5862d510f05600c179fe0b6062ebc67517bd402f5f092bb not found: ID does not exist" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.174625 4492 scope.go:117] "RemoveContainer" containerID="54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b" Feb 20 09:32:09 crc kubenswrapper[4492]: E0220 09:32:09.174971 4492 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b\": container with ID starting with 54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b not found: ID does not exist" containerID="54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.175023 4492 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b"} err="failed to get container status \"54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b\": rpc error: code = NotFound desc = could not find container \"54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b\": container with ID starting with 54f3c90d592891068a340a9e1e38e5c1c03dccd9cd1db4a0059dbcc72674ec8b not found: ID does not exist" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.311172 4492 patch_prober.go:28] interesting pod/machine-config-daemon-zrlsb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.311671 4492 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.311736 4492 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.312402 4492 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fbaac528affb82db28220a7361d65e6c7d8198f9422e8202431bdfff10da5453"} pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.312468 4492 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" podUID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerName="machine-config-daemon" containerID="cri-o://fbaac528affb82db28220a7361d65e6c7d8198f9422e8202431bdfff10da5453" gracePeriod=600 Feb 20 09:32:09 crc kubenswrapper[4492]: I0220 09:32:09.568803 4492 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1121159f-d018-464b-a012-afad30e73f00" path="/var/lib/kubelet/pods/1121159f-d018-464b-a012-afad30e73f00/volumes" Feb 20 09:32:10 crc kubenswrapper[4492]: I0220 09:32:10.098062 4492 generic.go:334] "Generic (PLEG): container finished" podID="f8755354-8581-4b5d-9b7e-6b7c53707f8c" containerID="fbaac528affb82db28220a7361d65e6c7d8198f9422e8202431bdfff10da5453" exitCode=0 Feb 20 09:32:10 crc kubenswrapper[4492]: I0220 09:32:10.098149 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerDied","Data":"fbaac528affb82db28220a7361d65e6c7d8198f9422e8202431bdfff10da5453"} Feb 20 09:32:10 crc kubenswrapper[4492]: I0220 09:32:10.098399 4492 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zrlsb" event={"ID":"f8755354-8581-4b5d-9b7e-6b7c53707f8c","Type":"ContainerStarted","Data":"b7d8fe28951f97adab9bcaf1a1f5b7628a590fce22c12114a5015ff2307c7238"} Feb 20 09:32:10 crc kubenswrapper[4492]: I0220 09:32:10.098426 4492 scope.go:117] "RemoveContainer" containerID="b1dc5c39aad26c0c31a11f8298847f9a14ac6859d0d60e93b504e8178355674a" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515146025166024453 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015146025167017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015146000513016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015146000513015450 5ustar corecore